Articles

404 ERROR: REQUEST COULD NOT BE FOUND

The page that you have requested could not be found at this time. We have provided you a list of related content below or you can use our site search to find the information that you are looking for.

AMD Confirms Stadia Will Run on Intel CPUs

As one of the world's most pervasive cloud service providers, Google is in a better position to launch a successful game streaming platform than almost anyone. The hardware they choose to use for the launch of their "Stadia" streaming service will undoubtedly influence future game streaming efforts, hence AMD's stock price shot through the roof when Google announced they were using AMD GPUs. However, PCGamesN writer Dave James noticed that Google was conspicuously silent when it came to Stadia's CPUs. They were happy to share clock speeds, cache numbers and the fact that they're using "custom" x86 chips, but they refused to confirm the vendor of the platform's CPU. Eventually, AMD reached out and said that "the Stadia platform is using custom AMD Radeon datacentre GPUs, not AMD CPUs." Barring any surprise announcements from VIA, that more or less confirms that Stadia will run on some sort of Intel CPU platform, but just why Google refused to mention Chipzilla by name remains a mystery. The author suggested that Intel might not want to associate themself with what might be a "doomed" venture. Maybe Google plans to switch to EPYC CPUs or an unannounced Intel server platform sometime in the future, or maybe they just don't think it's particularly relevant. Whatever the reason may be, I also find the omission to be curious, and look forward to seeing what happens with Stadia's hardware in the future.

A switch to AMD's EPYC processors has been mooted as a potential future step for Stadia, and Google's Phil Harrison told us himself that "we're just talking about Gen 1 at the moment, but there will be iterations on that technology over time," so there is some potential for a changing of the processor guard either before or after launch. Whatever the truth of the matter is I still find it beyond strange that no-one involved is talking about the Intel CPUs being used for Google Stadia, even if they're not necessarily doing anything that special with regards the innovative streaming service. Certainly the multi-GPU features on offer with the Radeon graphics cards warranted mention, but just a note on the specs slide alone could have still done good things for Intel.

Discussion
Posted by alphaatlas March 22, 2019 11:20 AM (CDT)

Intel Previews Processors and Graphics Software at GDC 2019

At their GDC 2019 conference, Intel confirmed that they'll launch 9th generation mobile processors in the 2nd quarter of 2019. While 9th generation H-series and Y-series "Ice Lake" parts recently showed up on the EEC website, Intel told PC World that these parts are based on 14nm Coffee Lake Silicon. The company also mentioned that one of their goals with his release is "longer battery life" for gamers and more casual users alike, and they're promoting their Wi-Fi 6 capable AX200 chip and 3D XPoint memory with the new chips Meanwhile, Intel also showed off a new software suite for their modern IGPs and (presumably) their future GPUs. The "Intel Graphics Command Center" is essentially their answer to Nvidia's GeForce Experience and AMD's Game Advisor, as it automatically scans your PC for supported games and applies the optimal settings for your current hardware. An "early access" version of the control panel is available on the Microsoft Store, and oddly enough, it says it was "released" on 11/26/2018. Unlike other app stores, the Microsoft Store doesn't log updates or list old changes, so it was presumably in some kind of closed alpha before being officially launched today.

We asked, you answered. You're tired of our 'old, boring, corporate-looking' Graphics Control Panel. We were too and we designed a completely new one from the ground up! We're incorporating the changes you - the gamers, home theater enthusiasts, professionals, and everyday tinkerers requested. Using a phased approach, we're rolling out something we're proud to share with you: introducing the Intel Graphics Command Center.

Discussion
Posted by alphaatlas March 21, 2019 9:54 AM (CDT)

Intel Shows Off Graphics Card Designs at GDC 2019

Intel reportedly unveiled some "early designs" of their upcoming discrete graphics cards at their GDC 2019 presentation. The graphics card in the first and 2nd slides they showed largely reassembles an Optane 905P SSD with a blower fan and a conspicuously short PCB. While the Xe's specs and performance levels are still unknown, to me, the short PCB suggests that Intel will use some kind of on-package memory with their upcoming GPU, or a relatively narrow GDDR memory bus at the very least. A shot of the back reveals a full backplate, as well as 3 DisplayPort outputs and one HDMI port. Finally, the last slide shows a card with a fan right on top of the graphics chip, which is something I haven't seen on a high-end reference card in some time.

Unfortunately, full specifications are still not yet available for Intel's upcoming graphics card. Real world performance is essentially completely unknown for now. As the year goes on, there is a good chance Intel may share some numbers given how eager the company is to make everyone aware that they have a major new product incoming.

Discussion
Posted by alphaatlas March 21, 2019 8:32 AM (CDT)

Intel Ice Lake Shows Up In EEC Database

Intel showed off a 10nm Ice Lake "client SoC" at CES this year, and revealed that it will use the "Sunny Cove" CPU architecture. While they gave a few details about the upcoming mobile chips and the core itself, we didn't hear much about Ice Lake in higher power parts. However, Twitter user and data-miner Komachi has once again found some unreleased hardware on the Eurasian Economic Commission's Online Portal. The first listing shows an "Idaville Ice Lake-D Pre-Alpha 85W Clear Linux Internal 32G Physical SDP," suggesting that Intel will brink the upcoming 10nm architecture to their (relatively) high power Xeon-D server chip lineup. Assuming the listing is accurate (as some other EEC listings have been,) this more or less confirms that Ice Lake won't be confined to the realm of low-power laptop chips.
Meanwhile, the next listing suggests that the low power "Ice Lake-Y" chips will have a "4+2" core config. Intel's current Amber Lake processors top out at 2 cores, so if I'm reading the listing right, it looks like ultra low power notebooks could get a core count boost next generation. There's also an Ice Lake-U "upgrade kit" listing with the same "4+2" core config. Discussion
Posted by alphaatlas March 19, 2019 9:48 AM (CDT)

Goodbye HardOCP - Hello Intel

We have some big changes happening here at HardOCP. Kyle Bennett will be taking on new challenges very soon with Intel working as its Director of Enthusiast Engagement.
Posted by Kyle March 19, 2019 6:30 AM (CDT)

Microsoft Announces Variable Rate Shading Support for DX12

Variable Rate Shading (VRS) is a powerful new API that gives the developers the ability to use GPUs more intelligently. Shaders are used to calculate the color of each pixel in a screen. Shading rate refers to the resolution at which these shaders are called (which is different from the overall screen resolution). A higher shading rate means better visual fidelity at the cost of using more GPU power. All pixels in a frame are affected by the game's shading rate. VSR allows developers to choose which areas of the frame are more important and increase the visual fidelity, or set parts of the frame to have lower fidelity and gain extra performance. Lowering the fidelity of parts of the scene can help low spec machines to run faster. There are two tiers of support for VRS. First of all the VRS API lets developers set the shading rate in 3 different ways: per draw, within a draw by using a screenspace image, or within a draw, per primitive. The hardware that can support per-draw VRS hardware are Tier 1. There's also a Tier 2, the hardware that can support both per-draw and within-draw variable rate shading. VRS support exists today on in-market NVIDIA hardware and on upcoming Intel hardware. AMD is rumored to be working on support for the feature.

For example, foveated rendering, rendering the most detail in the area where the user is paying attention, and gradually decreasing the shading rate outside this area to save on performance. In a first-person shooter, the user is likely paying most attention to their crosshairs, and not much attention to the far edges of the screen, making FPS games an ideal candidate for this technique. Another use case for a screenspace image is using an edge detection filter to determine the areas that need a higher shading rate, since edges are where aliasing happens. Once the locations of the edges are known, a developer can set the screenspace image based on that, shading the areas where the edges are with high detail, and reducing the shading rate in other areas of the screen.

Discussion
Posted by cageymaru March 18, 2019 7:07 PM (CDT)

Intel Delivers First Exascale Supercomputer to Argonne National Laboratory

Intel Corporation and Cray Inc. have announced that a Cray "Shasta" system will be the first U.S. exascale supercomputer. This $500 million Aurora supercomputer will be coming to the U.S. Department of Energy's Argonne National Laboratory in 2021 and will have a performance of one exaFLOP - a quintillion floating point operations per second. In addition, this system is designed to enable the convergence of traditional HPC, data analytics, and artificial intelligence -- at exascale. The program contract is valued at more than $100 million for Cray, one of the largest contracts in the company's history. The design of the Aurora system calls for 200 Shasta cabinets, Cray's software stack optimized for Intel architectures, Cray Slingshot interconnect, as well as next generation Intel technology innovations in compute processor, memory and storage technologies. Intel's Rajeeb Hazra detailed some of the futuristic technology coming to Aurora including a future generation Intel Xeon Scalable processor, the recently announced Intel Xe compute architecture, and Intel Optane DC persistent memory. "Today is an important day not only for the team of technologists and scientists who have come together to build our first exascale computer -- but also for all of us who are committed to American innovation and manufacturing," said Bob Swan, Intel CEO. "The convergence of AI and high-performance computing is an enormous opportunity to address some of the world's biggest challenges and an important catalyst for economic opportunity."

The Aurora system's exaFLOP of performance -- equal to a "quintillion" floating point computations per second -- combined with an ability to handle both traditional high-performance computing (HPC) and artificial intelligence (AI) will give researchers an unprecedented set of tools to address scientific problems at exascale. These breakthrough research projects range from developing extreme-scale cosmological simulations, discovering new approaches for drug response prediction and discovering materials for the creation of more efficient organic solar cells. The Aurora system will foster new scientific innovation and usher in new technological capabilities, furthering the United States' scientific leadership position globally.

Discussion
Posted by cageymaru March 18, 2019 3:24 PM (CDT)

Toyota Has a Curious Justification for Not Selling Any EVs

Toyota is one the biggest automakers, but it doesn’t sell a single fully electric vehicle in the United States. Why is that? According to vice president of research and development for Europe Gerald Killmann, it has all to do with battery production -- or the lack thereof. Toyota’s manageable supply can only allow for either 28,000 EVs or 1.5 million hybrid cars, and they’ve chosen the latter for being the more environmentally sound choice. Critics say the automaker is making excuses for terrible business decisions that have forced it to the back of the EV race.

...the calculation seems to assume that for every hybrid sold, a fully gasoline-powered car would be taken off the road. In reality, many Toyota hybrid buyers are replacing a Toyota hybrid. And, based on Toyota’s own revelation that they are losing Prius drivers to Tesla, it stands to reason that many Toyota hybrid drivers would jump at the opportunity to transition to an all-electric Toyota. Ultimately, Toyota's strategic decision to invest in gasoline-electric hybrids and bet on fuel cells in the long term is the reason that it isn't currently producing any electric cars.

Discussion
Posted by Megalith March 17, 2019 3:45 PM (CDT)

Win Some, Lose Some in Qualcomm vs Apple Saga

A jury in federal court in San Diego awarded Qualcomm $31 million in a patent infringement case against Apple. Qualcomm requested damages amounting to $1.41 per iPhone. The three patents that Apple allegedly violated were related to battery life in mobile phones. In another case, a U.S. federal judge issued a preliminary ruling favoring Apple related to patent royalty rebate payments. Factories that build Apple iPhones paid Qualcomm billions of dollars to use Qualcomm's technology. Apple agreed to reimburse the factories. "Qualcomm and Apple had a cooperation agreement under which Qualcomm would pay Apple a rebate on the iPhone patent payments if Apple agreed not to attack in court or with regulators." Apple filed a lawsuit after it says Qualcomm refused to issue $1 billion in patent royalty rebates. Qualcomm alleged that "Apple had broken the agreement by urging other smartphone makers to complain to regulators and making "false and misleading" statements to the Korean Fair Trade Commission." Apple said it was just responding to regulators.

The decision will not become final until after the trial in the case, which begins next month. And it is unlikely that Qualcomm will make a new payment to Apple. Apple's contract factories, which under normal circumstances would pay Qualcomm for patent royalties owed on iPhones, have already withheld the nearly $1 billion in payments to Qualcomm. "Apple has already offset the payment at issue under the agreement against royalties that were owed to Qualcomm," Qualcomm's Rosenberg told Reuters.

Discussion
Posted by cageymaru March 15, 2019 3:32 PM (CDT)

Intel's 5G Modems Will Allegedly Enter Mass Production in 2020

Ever since unveiling their 5G modems in 2017, Intel has been talking up 5G technology as loudly as they possibly can. However, facing stiff competition from rivals like Qualcomm, Intel more or less acknowledged that that their first generation modem won't be particularly competitive, and recently "made a strategic decision to pull in the launch of this [second generation] modem by half a year to deliver a leading 5G solution." Intel claimed they would introduce the more advanced XMM 8160 modem in the 2nd half of 2019, but Digitimes' industry sources think it won't be ready for mass production until 2020.

Intel is reportedly to begin working on engineering projects that will enable mass-production of 5G modem chips with its collaborative partners in the second quarter of 2019, according to sources from Taiwan's IC backend service providers... Intel is gearing up efforts to compete with Qualcomm, or even MediaTek, for 5G modem chip orders from Apple for its next-generation iPhone devices, the sources noted. However, judging from factors including heterogeneous integration, complexity of 5G modem chip design, and lengthy final test (FT) of relevant chips at packaging-level testing, it seems that Intel is unlikely to enter volume production of 5G modem chips until 2020, indicated the sources. Nevertheless, demand for Intel's modem chips for use in the Phone 8 and even iPhone 7 series will continue in the first half of 2019 as sales of the old-generation iPhones still remain robust, said the sources.

Discussion
Posted by alphaatlas March 15, 2019 9:58 AM (CDT)

AMD Radeon Software Adrenalin 2019 Edition 19.3.2 Supports DX12 on Windows 7

The AMD Radeon Software Adrenalin 2019 Edition 19.3.2 driver has been released and it is optimized for Tom Clancy's The Division 2 and Sid Meier's Civilization VI: Gathering Storm. AMD Radeon VII owners should see a 4% performance uplift in Sid Meier's Civilization VI: Gathering Storm when compared to the 19.2.3 driver. This driver brings DirectX 12 to Windows 7 for supported game titles. AMD supports more Vulkan extensions in this driver. Fixed issues include: Radeon ReLive for VR may sometimes fail to install during Radeon Software installation. Fan curve may fail to switch to manual mode after the manual toggle is switched when fan curve is still set to default behavior. Changes made in Radeon WattMan settings via Radeon Overlay may sometimes not save or take effect once Radeon Overlay is closed.

Known issues include: Rainbow Six Siege may experience intermittent corruption or flickering on some game textures during gameplay. DOTA2 VR may experience stutter on some HMD devices when using the Vulkan API. Mouse cursors may disappear or move out of the boundary of the top of a display on AMD Ryzen Mobile Processors with Radeon Vega Graphics. Performance metrics overlay and Radeon WattMan gauges may experience inaccurate fluctuating readings on AMD Radeon VII.

Discussion
Posted by cageymaru March 14, 2019 8:41 PM (CDT)

Intel Comet Lake Processors Could Pack 10 Cores

A recently updated file in coreboot's Github repository seemingly corroborates previous rumors that claim Intel's upcoming Comet Lake processors could pack up to 10 cores. The "report_platform.c" file contains references to various Comet Lake CPU + Graphics core configurations, including a "CometLake-S (10+2)" config.
The listing also suggests that Intel will launch low power, 6-core U-series parts for laptops, as well as a variety of other 2, 4, 6, and 8 core configs designed to supplement or replace the existing Coffee Lake and Whiskey Lake lineup. Discussion
Posted by alphaatlas March 14, 2019 11:05 AM (CDT)