Raevenlord
News Editor
- Joined
- Aug 12, 2016
- Messages
- 3,755 (1.23/day)
- Location
- Portugal
System Name | The Ryzening |
---|---|
Processor | AMD Ryzen 9 5900X |
Motherboard | MSI X570 MAG TOMAHAWK |
Cooling | Lian Li Galahad 360mm AIO |
Memory | 32 GB G.Skill Trident Z F4-3733 (4x 8 GB) |
Video Card(s) | Gigabyte RTX 3070 Ti |
Storage | Boot: Transcend MTE220S 2TB, Kintson A2000 1TB, Seagate Firewolf Pro 14 TB |
Display(s) | Acer Nitro VG270UP (1440p 144 Hz IPS) |
Case | Lian Li O11DX Dynamic White |
Audio Device(s) | iFi Audio Zen DAC |
Power Supply | Seasonic Focus+ 750 W |
Mouse | Cooler Master Masterkeys Lite L |
Keyboard | Cooler Master Masterkeys Lite L |
Software | Windows 10 x64 |
The potential impact of Artificial Intelligence (AI) has never been greater - but we'll only be successful if AI can deliver smarter and more intuitive answers. A key barrier to AI today is that natural data fed to a computer is largely unstructured and "noisy."
It's easy for humans to sort through natural data. For example: If you are driving a car on a residential street and see a ball roll in front of you, you would stop, assuming there is a small child not far behind that ball. Computers today don't do this. They are built to assist humans with precise productivity tasks. Making computers efficient at dealing with probabilities at scale is central to our ability to transform current systems and applications from advanced computational aids into intelligent partners for understanding and decision-making.
This is why probabilistic computing is one key component to AI and central to addressing these challenges. Probabilistic computing will allow future systems to comprehend and compute with uncertainties inherent in natural data, which will enable us to build computers capable of understanding, predicting and decision-making.
Today at Intel, we are observing an unprecedented growth of applications that rely on analysis of noisy natural data - different and even conflicting information. Such applications aim to assist humans with a higher level of intelligence and awareness about the environments in which they operate. Cutting through this noisy minefield is central to our ability to transform computers into intelligent partners that can understand and act on information with human-like fidelity.
Research into probabilistic computing is not a new area of study, but the improvements in high-performance computing and deep learning algorithms may lead probabilistic computing into a new era. In the next few years, we expect that research in probabilistic computing will lead to significant improvements in the reliability, security, serviceability and performance of AI systems, including hardware designed specifically for probabilistic computing. These advancements are critical to deploying applications into the real world - from smart homes to smart cities.
To accelerate our work in probabilistic computing, Intel is increasing its research investment in probabilistic computing and we are working with partners to pursue this goal.
Establishing the Intel Strategic Research Alliance for Probabilistic Computing
Realizing the full potential of probabilistic computing involves holistic integration of multiple levels in computing technology. Today, Intel underscored its commitment to integrated and collaborative implementation of emerging computing architectures and a sound ecosystem enablement strategy by issuing a call to the academic and start-up communities to partner with us to advance probabilistic computing from the lab to reality across these vectors: benchmark applications, adversarial attack mitigations, probabilistic frameworks and software and hardware optimization.
An Eye on What's Next
We are incredibly eager to see the proposals to advance probabilistic computing and to continue this research with the potential to raise the bar for what AI can help us achieve. Academic proposals are expected to be submitted by May 25th and among them we will select the best research teams.
We began this journey with research into neuromorphic computing - focusing on our understanding of the human brain and its associated computational processes. The start of the neuromorphic research community announced on March 1 is also on track and we are planning to continue to scale up our Loihi on the cloud to allow researchers access to cutting-edge hardware. We see a path to reach 100 billion synapses on a single system in 2019.
Furthermore, Intel has already been working to decode the brain and advance the next stage in neuroscience as part of our research partnership with Princeton University. We are looking forward to further understanding the flow of intelligence and decision-making through our probabilistic computing work.
View at TechPowerUp Main Site
It's easy for humans to sort through natural data. For example: If you are driving a car on a residential street and see a ball roll in front of you, you would stop, assuming there is a small child not far behind that ball. Computers today don't do this. They are built to assist humans with precise productivity tasks. Making computers efficient at dealing with probabilities at scale is central to our ability to transform current systems and applications from advanced computational aids into intelligent partners for understanding and decision-making.
This is why probabilistic computing is one key component to AI and central to addressing these challenges. Probabilistic computing will allow future systems to comprehend and compute with uncertainties inherent in natural data, which will enable us to build computers capable of understanding, predicting and decision-making.
Today at Intel, we are observing an unprecedented growth of applications that rely on analysis of noisy natural data - different and even conflicting information. Such applications aim to assist humans with a higher level of intelligence and awareness about the environments in which they operate. Cutting through this noisy minefield is central to our ability to transform computers into intelligent partners that can understand and act on information with human-like fidelity.
Research into probabilistic computing is not a new area of study, but the improvements in high-performance computing and deep learning algorithms may lead probabilistic computing into a new era. In the next few years, we expect that research in probabilistic computing will lead to significant improvements in the reliability, security, serviceability and performance of AI systems, including hardware designed specifically for probabilistic computing. These advancements are critical to deploying applications into the real world - from smart homes to smart cities.
To accelerate our work in probabilistic computing, Intel is increasing its research investment in probabilistic computing and we are working with partners to pursue this goal.
Establishing the Intel Strategic Research Alliance for Probabilistic Computing
Realizing the full potential of probabilistic computing involves holistic integration of multiple levels in computing technology. Today, Intel underscored its commitment to integrated and collaborative implementation of emerging computing architectures and a sound ecosystem enablement strategy by issuing a call to the academic and start-up communities to partner with us to advance probabilistic computing from the lab to reality across these vectors: benchmark applications, adversarial attack mitigations, probabilistic frameworks and software and hardware optimization.
An Eye on What's Next
We are incredibly eager to see the proposals to advance probabilistic computing and to continue this research with the potential to raise the bar for what AI can help us achieve. Academic proposals are expected to be submitted by May 25th and among them we will select the best research teams.
We began this journey with research into neuromorphic computing - focusing on our understanding of the human brain and its associated computational processes. The start of the neuromorphic research community announced on March 1 is also on track and we are planning to continue to scale up our Loihi on the cloud to allow researchers access to cutting-edge hardware. We see a path to reach 100 billion synapses on a single system in 2019.
Furthermore, Intel has already been working to decode the brain and advance the next stage in neuroscience as part of our research partnership with Princeton University. We are looking forward to further understanding the flow of intelligence and decision-making through our probabilistic computing work.
View at TechPowerUp Main Site