Currently showcasing 12 feeds. More coming soon!
OpenAI has released a new series of AI models, called OpenAI o1-Preview, that are designed to improve reasoning and problem-solving capabilities.
These models have been trained to refine their thinking processes, try different methods, and recognize mistakes before providing an answer.
In tests, the models performed comparably to Ph.D. students in tasks related to physics, chemistry, biology, mathematics, and coding.
Apple's new iPhone 16 and 16 Pro models, with advanced AI capabilities, are expected to drive a boom in sales and start an iPhone "supercycle," according to some analysts. The AI features, such as improved photo editing, custom emoji generation, and a more natural Siri voice assistant, are seen as a major draw for customers. However, other analysts are skeptical of the supercycle prediction, noting that the AI features may not be compelling enough to drive large numbers of early upgrades.
This article provides a tutorial on creating a pronunciation assessment app.
The app's goal is to help users improve their pronunciation by allowing them to enter a word, record their voice, and receive a score based on the API's analysis.
Knowledge of JavaScript and Vue.js 3 is recommended to follow the tutorial.
AI researchers at Tsinghua University and Zhipu AI have developed a large language model (LLM) called LongWriter that can generate text output of up to 10,000 words.
Existing LLMs are typically limited to generating short outputs of around 2,000 words because they are trained on short documents. LongWriter overcomes this limitation by training on longer documents.
The researchers have made the code for LongWriter open-source, allowing others to build upon their work. They also posted a video demonstrating LongWriter generating a 10,000-word tourist guide.
Roeland Decorte, inspired by his experience in codebreaking, developed a smartphone app that uses AI to listen for signs of disease hidden in a person's pulse.
Decorte's company, Decorte Future Industries, is at the forefront of an audio-powered revolution in healthcare, using algorithms to interpret the body's faint signals and diagnose various conditions, including heart problems and stomach cancer.
By using just a microphone, Decorte's technology can provide accurate readings at home, eliminating the need for multiple apps and hardware solutions for different conditions.
Researchers have ranked AI models based on their risk level, revealing a wide range of behaviors and rule-breaking tendencies among these models.
Regulations may need to be tightened to address the legal, ethical, and regulatory compliance issues associated with AI models.
Government rules and guidelines on AI are found to be less comprehensive than companies' policies, suggesting room for improvement and the need for stricter regulations.
1. MIT researchers have developed SigLLM, a framework that uses large language models (LLMs) to detect anomalies in time-series data without the need for training or fine-tuning. LLMs have the potential to be more efficient and cost-effective than deep-learning models for anomaly detection tasks.
2. The researchers found that LLMs can convert time-series data into text-based inputs that the models can process. They developed two approaches, Prompter and Detector, that use LLMs to locate anomalous values and predict future values, respectively.
3. While LLMs did not outperform state-of-the-art deep learning models, they showed promise in anomaly detection and could be used to flag potential problems in equipment such as wind turbines or satellites before they occur. Future work will focus on improving LLM performance and understanding their performance in anomaly detection tasks.
Scientists from MIT's Computer Science and Artificial Intelligence Laboratory (CSAIL) have developed a method to train robots in simulated environments created using "digital twins" of real-world spaces. Users can scan a physical environment using a smartphone app and make necessary adjustments to create a digital replica. Robots can then train in the simulated environment, which accelerates the learning process compared to training in the real world.
The approach eliminates the need for extensive reward engineering and can create strong policies for various tasks, improving performance even in environments with disturbances and distractions. The researchers tested the system's performance in controlled lab settings and real-world environments and found that it outperformed traditional imitation-learning methods, especially in situations with visual distractions or physical disruptions.
RialTo, the robot training system developed by the researchers, currently takes three days to be fully trained, but the team is working on improving the training process and the model's adaptability to new environments. The researchers presented their work at the Robotics Science and Systems (RSS) conference.
MIT researchers are using machine learning to accurately measure the atomic patterns in metals, known as short-range order (SRO), which is crucial for designing custom materials. The goal is to use SRO as a tool to tailor material properties in high-entropy alloys, which have complex compositions and superior properties. Machine learning models are used to identify chemical motifs and quantify SRO, providing a more comprehensive understanding of these materials.
Researchers from MIT and other institutions have developed a new machine-learning framework that can predict phonon dispersion relations, which are key to understanding how heat moves through materials, up to 1,000 times faster than other AI-based techniques. The method, called the virtual node graph neural network (VGNN), uses flexible virtual nodes to represent phonons in a crystal structure, allowing for more efficient and accurate predictions. This technology could help engineers design more efficient energy-conversion systems and faster microelectronic devices by reducing waste heat.
The VGNN method is capable of predicting phonon dispersion relations for a few thousand materials in just a few seconds using a personal computer, significantly improving the efficiency and speed of these calculations compared to traditional methods. The researchers propose three different versions of VGNNs with increasing complexity, which can be used to predict phonons directly from a material's atomic coordinates. The technique can also be extended to predict other high-dimensional quantities, such as optical and magnetic properties.
The VGNN method offers comparable or even better accuracy compared to other AI-based techniques, making it a promising tool for predicting thermal properties of materials. By enabling faster and more efficient predictions of phonon dispersion relations, this method could help in the design of energy generation systems and microelectronics that produce more power and
A new technique has been developed by MIT researchers to assess the reliability of foundation models, which are large pretrained deep-learning models used in AI applications. The technique involves training a set of models that are slightly different from one another and assessing the consistency of their representations of the same test data point. This technique can be used to determine if a model is reliable for a specific task, without needing to test it on real-world data.
The technique outperformed state-of-the-art baseline methods in capturing the reliability of foundation models across various classification tasks. It can also be used to rank models based on their reliability scores, allowing users to select the best model for their needs.
The researchers used an ensemble approach, training multiple models with shared properties but slight differences. They used an idea called neighborhood consistency to compare the abstract representations outputted by the models and estimate their reliability. This approach aligns the models' representation spaces by using neighboring points as anchors. The technique was found to be more consistent and robust than other methods, even with challenging test points. However, training an ensemble of large foundation models can be computationally expensive, so the researchers plan to explore more efficient methods in the future.
Researchers from MIT and the MIT-IBM Watson AI Lab have developed a technique to estimate the reliability of foundation models, which are massive deep-learning models pretrained on general-purpose, unlabeled data. The technique involves training a set of foundation models that are slightly different from one another and comparing the consistency of the representations each model learns about the same test data point. The technique outperformed state-of-the-art baseline methods in capturing the reliability of foundation models on a variety of classification tasks.
The technique can be used to assess the reliability of foundation models before they are deployed to a specific task, which is particularly useful in safety-critical situations where incorrect or misleading information could have serious consequences. It also enables users to choose the most reliable model for their task and does not require testing on a real-world dataset.
One limitation of the technique is that it requires training an ensemble of large foundation models, which is computationally expensive. Future work will focus on finding more efficient ways to build multiple models.
1. OpenAI whistleblowers have filed a complaint with the SEC, alleging that the company restricted workers from speaking out about the risks of its AI technology.
2. The whistleblowers are asking the SEC to investigate OpenAI's non-disclosure agreements and enforce rules against discouraging employees from raising concerns with regulators.
3. U.S. Senator Chuck Grassley has called for changes to OpenAI's policies and practices, stating that they have a chilling effect on whistleblowers' right to speak up.
Researchers at Carnegie Mellon University have identified six downstream harms caused by voice assistant errors for users with multicultural backgrounds, including emotional, cultural, and relational harm. These harms can be experienced as microaggressions and have a negative impact on self-esteem and sense of belonging. The researchers suggest strategies such as blame redirection and increasing cultural sensitivity in voice technologies to reduce these harms.
Voice assistants that are trained on datasets that predominantly represent white Americans are more likely to misinterpret and misunderstand Black speakers or people with accents or dialects that differ from standard American. This has led to harmful consequences for users with multicultural backgrounds, including higher self-consciousness and negative views of technology. The ultimate solution is to eliminate bias in voice technologies, but this is a challenging task that requires creating representative datasets.
One communication repair strategy suggested by the researchers is blame redirection, where the voice assistant explains the error without blaming the user. They also recommend increasing the database of proper nouns to address misrecognition of non-Anglo names. Another approach is to include affirmations in voice assistant conversations to protect the user's identity. However, brevity is essential in these interventions to maintain efficiency and hands-free use.
NVIDIA researchers will present advancements in simulation and generative AI at the SIGGRAPH conference, focusing on diffusion models for visual generative AI, physics-based simulation, and realistic AI-powered rendering.
The research includes innovations in generating consistent imagery for storytelling, real-time texture painting on 3D meshes, simulating complex human motions based on text prompts, and modeling the behavior of objects in different environments.
NVIDIA-authored papers also introduce techniques for faster modeling of visible light, simulating diffraction effects, improving the quality of path tracing algorithms, and creating multipurpose AI tools for 3D representation and design.
PlayFi has launched the PlayFi Airdrop Platform, an AI-powered data network and blockchain designed specifically for the gaming industry.
The platform serves as a central hub for earning points and engaging with the PlayFi community.
Through the platform, users can transform their interaction with live content and participate in the PlayFi ecosystem using $PLAY tokens.
1. Researchers have developed a new AI model that can accurately predict if a person will develop dementia up to five years in advance. The model uses data from electronic health records and can provide early diagnosis and intervention for patients at risk.
2. The AI model was trained on a large dataset of electronic health records from over 900,000 patients. It was able to accurately identify individuals who would later develop dementia, including Alzheimer's disease, with an accuracy of 94%.
3. Early detection of dementia is crucial for effective treatment and intervention. With the use of AI models like this, healthcare professionals can potentially identify at-risk patients earlier, leading to improved outcomes and quality of life for those affected by the disease.
Eric Evans, director of MIT Lincoln Laboratory, has been awarded the Department of Defense Medal for Distinguished Public Service for his leadership and contributions to national security. Evans has advised multiple defense secretaries and secured funding for new facilities and test ranges. He will be stepping down as director but will continue to work with DoD leaders as a professor of practice at MIT.
MIT and Meta researchers have developed a computer vision technique called PlatoNeRF that can create accurate 3D reconstructions of scenes using images from a single camera position. By combining lidar technology with machine learning, PlatoNeRF can model the geometry of hidden objects and accurately reconstruct scenes, even in challenging lighting conditions. This technique could have applications in autonomous vehicles, AR/VR headsets, and warehouse robots.