Math

First Shape Found That Can't Pass Through Itself (quantamagazine.org) 35

Mathematicians have identified the first shape that cannot pass through itself. Jakob Steininger and Sergey Yurkevich described the Noperthedron in a paper posted online in August. The shape has 90 vertices and 152 faces. The discovery resolves a question that began in the late 1600s when Prince Rupert of the Rhine won a bet by proving one cube could slide through a tunnel bored through another. Mathematician John Wallis confirmed this mathematically in 1693.

The property became known as the Rupert property. In 1968, Christoph Scriba proved the tetrahedron and octahedron also possess this quality. Over the past decade, researchers found Rupert tunnels through many symmetric polyhedra, including the dodecahedron and icosahedron. Mathematicians had conjectured every convex polyhedron would have the Rupert property. Steininger and Yurkevich divided the space of possible orientations into approximately 18 million blocks and tested each. None produced a passage. The Noperthedron consists of 150 triangles and two regular 15-sided polygons.
The Military

Sweden's Crowd-Forecasting Platform 'Glimt' Helps Ukraine Make Wartime Predictions (france24.com) 20

alternative_right shares a report from France 24: [Sweden's] latest contribution to the war effort is Glimt, an innovative project launched by the Swedish Defence Research Agency (FOI) earlier this year. Glimt is an open platform that relies on the theory of "crowd forecasting": a method of making predictions based on surveying a large and diverse group of people and taking an average. "Glimt" is a Swedish word for "a glimpse" or "a sudden insight." The theory posits that the average of all collected predictions produces correct results with "uncanny accuracy," according to the Glimt website. Such "collective intelligence" is used today for everything from election results to extreme weather events, Glimt said. [...]

Group forecasting allows for a broad collection of information while avoiding the cognitive bias that often characterizes intelligence services. Each forecaster collects and analyses the available information differently to reach the most probable scenario and can add a short comment to explain their reasoning. The platform also encourages discussion between members so they can compare arguments and alter their positions. Available in Swedish, French and English, the platform currently has 20,000 registered users; each question attracts an average of 500 forecasters. Their predictions are later sent to statistical algorithms that cross-reference data, particularly the relevance of the answers they provided. The most reliable users will have a stronger influence on the results; this reinforces the reliability of collective intelligence.
"We used this method and research, and we suggested to the Ukrainians that it could improve their understanding of the world and its evolution," said Ivar Ekman, an analyst for the Swedish Defence Research Agency and program director for Glimt. "If you have a large group of people, you can achieve great accuracy in assessing future events. Research has shown that professional analysts don't necessarily have a better capacity in this domain than other people."
AI

OpenAI's 'Embarrassing' Math (techcrunch.com) 41

An anonymous reader writes: "Hoisted by their own GPTards." That's how Meta's Chief AI Scientist Yann LeCun described the blowback after OpenAI researchers did a victory lap over GPT-5's supposed math breakthroughs. Google DeepMind CEO Demis Hassabis added, "this is embarrassing." The Decoder reports that in a since-deleted tweet, OpenAI VP Kevin Weil declared that "GPT-5 found solutions to 10 (!) previously unsolved Erdos problems and made progress on 11 others." ("Erdos problems" are famous conjectures posed by mathematician Paul Erdos.)

However, mathematician Thomas Bloom, who maintains the Erdos Problems website, said Weil's post was "a dramatic misrepresentation" -- while these problems were indeed listed as "open" on Bloom's website, he said that only means, "I personally am unaware of a paper which solves it." In other words, it's not accurate to claim GPT-5 was able to solve previously unsolved problems. Instead, Bloom wrote, "GPT-5 found references, which solved these problems, that I personally was unaware of."

Space

'How We Sharpened the James Webb Telescope's Vision From a Million Kilometers Away' (theconversation.com) 18

The James Webb Space Telescope gets its highest resolution with the aperture masking interferometer (or AMI), "a tiny piece of precisely machined metal that slots into one of the telescope's cameras," according to a new article by Benjamin Pope, an associated math professor at Macquarie University.

"We can finally present its first successful observations of stars, planets, moons and even black hole jets." [AMI] was put on Webb to diagnose and measure any blur in its images. Even nanometres of distortion in Webb's 18 hexagonal primary mirrors and many internal surfaces will blur the images enough to hinder the study of planets or black holes, where sensitivity and resolution are key. AMI filters the light with a carefully structured pattern of holes in a simple metal plate, to make it much easier to tell if there are any optical misalignments. We wanted to use this mode to observe the birth places of planets, as well as material being sucked into black holes. But before any of this, AMI showed Webb wasn't working entirely as hoped.

At very fine resolution — at the level of individual pixels — all the images were slightly blurry due to an electronic effect: brighter pixels leaking into their darker neighbours. This is not a mistake or flaw, but a fundamental feature of infrared cameras that turned out to be unexpectedly serious for Webb. This was a dealbreaker for seeing distant planets many thousands of times fainter than their stars a few pixels away: my colleagues quickly showed that its limits were more than ten times worse than hoped. So, we set out to correct it...

We built a computer model to simulate AMI's optical physics, with flexibility about the shapes of the mirrors and apertures and about the colours of the stars. We connected this to a machine learning model to represent the electronics with an "effective detector model" — where we only care about how well it can reproduce the data, not about why. After training and validation on some test stars, this setup allowed us to calculate and undo the blur in other data, restoring AMI to full function. It doesn't change what Webb does in space, but rather corrects the data during processing. It worked beautifully — the star HD 206893 hosts a faint planet and the reddest-known brown dwarf (an object between a star and a planet). They were known but out of reach with Webb before applying this correction. Now, both little dots popped out clearly in our new maps of the system... With the new correction, we brought Jupiter's moon Io into focus, clearly tracking its volcanoes as it rotates over an hour-long timelapse.

"This correction has opened the door to using AMI to prospect for unknown planets at previously impossible resolutions and sensitivities..." the article points out.

"Our results on painstakingly testing and enhancing AMI are now released on the open-access archive arXiv in a pair of papers."

Thanks to long-time Slashdot reader schwit1 for sharing the article.
Education

South Korea Abandons AI Textbooks After Four-Month Trial (restofworld.org) 27

South Korea's government has stripped AI-powered textbooks of their official status after a single semester of use. The textbooks were introduced in March for math, English, and computer science classes as a flagship initiative under former President Yoon Suk Yeol. Students and teachers complained about technical problems, factual inaccuracies, and increased workload.

The government spent more than 1.2 trillion won ($850 million) on the program. Publishers invested around 800 billion won ($567 million). The textbooks were reclassified as supplementary material. Adoption rates dropped from 37% in the first semester to 19% in September. Only 2,095 schools now use them, about half the number from earlier in the year.
Math

The Numbers Six and Seven Are Making Life Hell for Math Teachers (msn.com) 165

Math teachers across American schools are contending with a classroom disruption that has proven impossible to contain. The numbers six and seven now trigger instant pandemonium among students. They scream the phrase and perform a palms-up seesaw hand gesture whenever the numbers appear in equations or instructions.

Teachers have begun avoiding breaking students into groups of six or seven or asking them to turn to page 67. The meme has no meaning, reports WSJ. That absence of meaning is the point. The phenomenon traces back to late last year when Philadelphia rapper Skrilla released "Doot Doot (6 7)," a song referencing 67th street where his friends grew up. The phrase spiraled into youth culture in March through a viral video of a boy with forward-swept hair lurching toward a camera to deliver an animated "six seven." Skrilla is now touring venues where audiences wait for the six-seven line. Some teachers have attempted to neutralize the meme by saying it themselves.
Cellphones

More Screen Time Linked To Lower Test Scores For Elementary Students (www.cbc.ca) 46

An anonymous reader quotes a report from CBC News: The study by a team from Toronto's Hospital for Sick Children (also known as Sick Kids) and St. Michael's Hospital was published in the Journal of the American Medical Association. It found that children who spent more time on screens before age eight scored lower on standardized tests. Child psychiatry researchers say handing kids digital devices, like iPads, every time they have a tantrum could lead to future issues. One new study links too much screen time to emotional and anger management problems.

The study followed more than 3,000 kids in Ontario over a 15 year span from 2008 to 2023, tracking how much time they spent watching TV or DVDs, playing video games, using the computer or playing on handheld devices like iPads, as reported by their parents. That data was compared to their EQAO standardized test scores, which are used to assess the reading and math skills of kids across Ontario in grades 3 and 6. The findings point to a "significant association," between screen use and lower test scores, according to Dr. Catherine Birken, a pediatrician and senior scientist at Sick Kids and lead author of the study.

"For each additional hour of screen use, there was approximately a 10 percent lower odds of meeting standards in both reading and mathematics ... in Grade 3 and mathematics in Grade 6," said Dr. Catherine Birken, a pediatrician and senior scientist at Sick Kids and lead author of the study, in an interview with CBC News. The study didn't differentiate between different types of screen time -- for example, whether a child was playing a game on their iPad versus FaceTiming a relative in another city, or watching an educational video. It was also an observational study that relied on parents answering questionnaires about how much time their kids spent in front of screens. The study authors note that this means the research can't be taken as definitive proof that screen time causes lower grades, just that the two things tend to go hand in hand.

AI

Mira Murati's Stealth AI Lab Launches Its First Product (wired.com) 33

An anonymous reader quotes a report from Wired: Thinking Machines Lab,a heavily funded startup cofounded by prominent researchers from OpenAI, has revealed its first product -- a tool called Tinker that automates the creation of custom frontier AI models. "We believe [Tinker] will help empower researchers and developers to experiment with models and will make frontier capabilities much more accessible to all people," said Mira Murati, cofounder and CEO of Thinking Machines, in an interview with WIRED ahead of the announcement.

Big companies and academic labs already fine-tune open source AI models to create new variants that are optimized for specific tasks, like solving math problems, drafting legal agreements, or answering medical questions. Typically, this work involves acquiring and managing clusters of GPUs and using various software tools to ensure that large-scale training runs are stable and efficient. Tinker promises to allow more businesses, researchers, and even hobbyists to fine-tune their own AI models by automating much of this work.

Essentially, the team is betting that helping people fine-tune frontier models will be the next big thing in AI. And there's reason to believe they might be right. Thinking Machines Lab is helmed by researchers who played a core role in the creation of ChatGPT. And, compared to similar tools on the market, Tinker is more powerful and user friendly, according to beta testers I spoke with. Murati says that Thinking Machines Lab hopes to demystify the work involved in tuning the world's most powerful AI models and make it possible for more people to explore the outer limits of AI. "We're making what is otherwise a frontier capability accessible to all, and that is completely game-changing," she says. "There are a ton of smart people out there, and we need as many smart people as possible to do frontier AI research."
"There's a bunch of secret magic, but we give people full control over the training loop," OpenAI veteran John Schulman says. "We abstract away the distributed training details, but we still give people full control over the data and the algorithms."
Science

Jane Goodall, Famed Primatologist and Conservationist, Dies At 91 (go.com) 26

Jane Goodall, world-renowned primatologist, anthropologist, and conservationist, has died at the age of 91 while on a speaking tour in California. The British primatologist's "discoveries as an ethologist revolutionized science, and she was a tireless advocate for the protection and restoration of our natural world," according to the institute she founded. From a report: Goodall was only 26 years old when she first traveled to Tanzania and began her important research on chimpanzees in the wild. Throughout her study of the species, Goodall proved that primates display an array of similar behaviors to humans, such as the ability to develop individual personalities and make and use their own tools. Among the most surprising discoveries Goodall made was "how like us" the chimpanzees are, she told ABC News in 2020.

"Their behavior, with their gestures, kissing, embracing, holding hands and patting on the back," she said. "... The fact that they can actually be violent and brutal and have a kind of war, but also loving an altruistic." That discovery is considered one of the great achievements of 20th-century scholarship, according to the Jane Goodall Institute. [...] Goodall's research garnered both scientific honors and mainstream fame, and she was credited with paving the way for a rise in women pursuing careers in STEM (science, technology, engineering and math) over the years. The number of women in STEM has increased from 7% to 26% in the six last decades, according to The Jane Goodall Institute, which cited census information from 1970 to 2011.

In 1991, she also founded Roots & Shoots, a global humanitarian and environmental program for young people. She was named a United Nations Messenger of Peace in April 2002. The anthropologist continued to lend her voice to environmental causes well into her 80s and 90s. In 2019, Goodall acknowledged the climate crisis and the importance of mitigating further warming, telling ABC News that the planet is "imperiled." "We are definitely at a point where we need to make something happen," she said.

"We are imperiled. We have a window of time. I'm fairly sure we do. But, we've got to take action." Goodall even partnered with Apple in 2022 to encourage customers to recycle their devices to reduce individual carbon footprint and cut down on unnecessary mineral mining around the world. "Yes, people need to make money, but it is possible to make money without destroying the planet," Goodall told ABC News at the time. "We've gone so far in destroying the planet that it's shocking."

Government

Should Salesforce's Tableau Be Granted a Patent On 'Visualizing Hierarchical Data'? 72

Long-time Slashdot reader theodp says America's Patent and Trademark Office (USPTO) has granted a patent to Tableau (Salesforce's visual analytics platform) — for a patent covering "Data Processing For Visualizing Hierarchical Data": "A provided data model may include a tree specification that declares parent-child relationships between objects in the data model. In response to a query associated with objects in the data model: employing the parent-child relationships to determine a tree that includes parent objects and child objects from the objects based on the parent-child relationships; determining a root object based on the query and the tree; traversing the tree from the root object to visit the child objects in the tree; determining partial results based on characteristics of the visited child objects such that the partial results are stored in an intermediate table; and providing a response to the query that includes values based on the intermediate table and the partial results."

A set of 15 simple drawings is provided to support the legal and tech gobbledygook of the invention claims. A person can have a manager, Tableau explains in Figures 5-6 of its accompanying drawings, and that manager can also manage and be managed by other people. Not only that, Tableau illustrates in Figures 7-10 that computers can be used to count how many people report to a manager. How does this magic work, you ask? Well, you "generate [a] tree" [Fig. 13] and "traverse a tree" [Fig. 15], Tableau explains. But wait, there's more — you can also display the people who report to a manager in multi-level or nested pie charts (aka Sunburst charts), Tableau demonstrates in Fig. 11.

Interestingly, Tableau released a "pre-Beta" Sunburst chart type in late April 2023 but yanked it at the end of June 2023 (others have long-supported Sunburst charts, including Plotly). So, do you think Tableau should be awarded a patent in 2025 on a concept that has roots in circa-1921 Sunburst charts and tree algorithms taught to first-year CS students in circa-1975 Data Structures courses?
United States

Decline in K-12 National Reading, Math, Science Scores Probed By US Senate Panel (newhampshirebulletin.com) 144

Just days after federal data revealed average reading, math and science scores dropped among certain grades since before the coronavirus pandemic, a U.S. Senate panel on Thursday picked apart the root causes and methods for students' academic improvement. From a report: The hearing in the Senate Committee on Health, Education, Labor and Pensions centered on the "state of K-12 education" -- which GOP members on the committee described as "troubling" -- in light of recent data from the National Assessment of Educational Progress, or NAEP.

NAEP, regarded as the gold standard for tracking students' academic performance, showed that average science scores for eighth-graders decreased by 4 points since before the pandemic, in 2019. Average math and reading scores for 12th-graders also fell 3 points between 2019 and 2024. The assessments were administered between January and March of 2024. Results also showed that just one-third of 12th-graders are considered academically prepared for college in math -- a drop from 37% in 2019.

The committee's chair, Sen. Bill Cassidy, said "it should concern us that children's reading, math and science scores have yet to recover to pre-pandemic levels." The Louisiana Republican added that "success in education is not determined by how much we spend, but by who makes the decision and how wisely resources are directed," and "when states and local communities are empowered to tailor solutions to meet the unique needs of students, innovation follows." On the other hand, Sen. Bernie Sanders, ranking member of the panel, said that "while we focus on education -- as important as that is -- we also have to focus on the conditions under which our children are living."

AI

UAE Lab Releases Open-Source Model to Rival China's DeepSeek (gizmodo.com) 43

"The United Arab Emirates wants to compete with the U.S. and China in AI," writes Gizmodo, "and a new open source model may be its strongest contender yet.

"An Emirati AI lab called the Institute of Foundation Models (IFM) released K2 Think on Tuesday, a model that researchers say rivals OpenAI's ChatGPT and China's DeepSeek in standard benchmark tests." "With just 32 billion parameters, it outperforms flagship reasoning models that are 20x larger," the lab wrote in a press release on Tuesday. DeepSeek's R1 has 671 billion parameters, though only 37 billion are active. Meta's latest Llama 4 models range from 17 billion to 288 billion active parameters. OpenAI doesn't share parameter information. OpenAI doesn't share parameter information.

Researchers also claim that K2 Think leads "all open-source models in math performance" across several benchmarks. The model is intended to be more focused on math, coding, and scientific research than most other AI chatbots. The Emirati lab's selling point for the model is similar to DeepSeek's strategy that disrupted the AI market earlier this year: optimized efficiency that will have better or the same computing power at a lower cost...

The lab is also aiming to be transparent in everything, "open-sourcing not just models but entire development processes" that provide "researchers with complete materials including training code, datasets, and model checkpoints," IFM said in a press release from May.

The UAE and other Arab countries are investing in AI to try reducing their economic dependence on fossil fuels, the article points out.
Education

US High School Students Lose Ground In Math and Reading, Continuing Yearslong Decline (apnews.com) 215

The latest National Assessment of Educational Progress shows U.S. high school seniors' math and reading scores at their lowest in decades, with nearly half failing to reach basic proficiency in math and one-third below basic in reading. The Associated Press reports: A decade-long slide in high schoolers' reading and math performance persisted during the COVID-19 pandemic, with 12th graders' scores dropping to their lowest level in more than 20 years, according to results released Tuesday from an exam known as the nation's report card. Eighth-grade students also lost significant ground in science skills, according to the results from the National Assessment of Education Progress.

The assessments were the first since the pandemic for eighth graders in science and 12th graders in reading and math. They reflect a downward drift across grade levels and subject areas in previous releases from NAEP, which is considered one of the best gauges of the academic progress of U.S. schools. "Scores for our lowest-performing students are at historic lows," said Matthew Soldner, the acting commissioner of the National Center for Education Statistics. "These results should galvanize all of us to take concerted and focused action to accelerate student learning." [...]

In reading, the average score in 2024 was the lowest score in the history of the assessment, which began in 1992. Thirty-two percent of high school seniors scored below "basic," meaning they were not able to find details in a text to help them understand its meaning. In math, the average score in 2024 was the lowest since 2005, when the assessment framework changed significantly. On the test, 45% of high school seniors scored below "basic" achievement, the highest percentage since 2005. Only 33% of high school seniors were considered academically prepared for college-level math courses, a decline from 37% in 2019.

Education

Dumbing Down the SAT Bodes Poorly for Education (bloomberg.com) 115

The SAT is billed as "a great way to find out how prepared students are for college." If that's true, recent changes to its format offer an unflattering assessment of the country's aspiring scholars, Bloomberg's editorial board wrote Wednesday. From the piece: [...] Then the pandemic hit. As in-person exams became impractical, hundreds of schools dropped their testing requirements. The SAT and its main competitor, the ACT, lost millions of dollars in revenue. Although both recently started offering digital options, schools have been slow to reinstate their requirements. Today, more than 80% of schools remain test-optional.

"If students are deciding to take a test," as one College Board executive put it, "how do we make the SAT the one they want to take?" To anyone familiar with American teenagers, the company's answer should come as no surprise: Make the test easier. The newly digitized format allows a calculator for the entire math section and drastically cuts reading comprehension. Gone are the 500- to 750-word passages about which students would answer a series of questions. Instead, test takers read 25- to 150-word excerpts -- about the length of a social media post -- and answer a single question about each.

[...] An effort by the College Board to reemphasize the benefits of deep reading -- for critical thinking, for self-reflection, for learning of all kinds -- might go a long way toward restoring some balance. It should build on efforts to incorporate college prep into school curricula, work with districts to develop coursework that builds reading stamina for all test takers, and consider reducing the cost of its subject-specific Advanced Placement exams that continue to test these skills (now $99), in line with the SAT ($68). Schools, for their part, should recommit to teaching books in their entirety.

Transportation

New Study Proves EVs Are Always Cleaner Than Gas Cars (thedrive.com) 195

An anonymous reader shares a report: It's broadly understood that electric vehicles are more environmentally friendly than their counterparts that burn only gasoline. And yes -- that includes the impact of manufacturing batteries and generating power to charge them. But even then, such generalizations gloss over specifics, like which EVs are especially eco-friendly, not to mention where. The efficiency of an electric car varies greatly depending on ambient temperature, which is less compromising for gas-burning vehicles.

We now have the data and math to answer these questions, courtesy of the University of Michigan. Last week, researchers there released a study along with a calculator that allows users to compare the lifetime difference in greenhouse gas emissions of various vehicle types and powertrains from "cradle to grave," as they say. That includes vehicle production and disposal, as well as use-phase emissions from "driving and upstream fuel production and/or electricity generation," per the university itself.

What's more, these calculations can be skewed by where you live. So, if I punch in my location of Bucks County, Pennsylvania, I can see that my generic, pure-ICE "compact sedan" emits 309 grams of carbon dioxide equivalent (gCO2e) per mile. A compact hybrid would emit 20% less; a plug-in hybrid, 44% less; and an EV with a 200-mile range, a whopping 63% less. And, if I moved to Phoenix, the gains would be even larger by switching to pure electric, to the tune of a 79% reduced carbon impact.

United States

The US Population Could Shrink in 2025, For the First Time Ever (derekthompson.org) 125

An anonymous reader shares a report: The United States is on the precipice of a historic, if dubious, achievement. If current trends hold, 2025 could be the first year on record in which the US population actually shrinks.

The math is straightforward. Population growth has two sources: natural increase (births minus deaths) and net immigration (arrivals minus departures). Last year, births outnumbered deaths by 519,000 people. That means any decline in net immigration in excess of half a million could push the U.S. into population decline. A recent analysis of Census data by the Pew Research Center found that between January and June, the US foreign-born population fell for the first time in decades by more than one million. While some economists have questioned the report, a separate analysis by the American Enterprise Institute predicted that net migration in 2025 could be as low as negative 525,000. In either case, annual population growth this year could easily turn negative.

AI

Harvard Dropouts To Launch 'Always On' AI Smart Glasses That Listen, Record Every Conversation 68

Two Harvard dropouts are launching Halo X, a $249 pair of AI-powered smart glasses that continuously listen, record, and transcribe conversations while displaying real-time information to the wearer. "Our goal is to make glasses that make you super intelligent the moment you put them on," said AnhPhu Nguyen, co-founder of Halo. Co-founder Caine Ardayfio said the glasses "give you infinite memory."

"The AI listens to every conversation you have and uses that knowledge to tell you what to say ... kinda like IRL Cluely," Ardayfio told TechCrunch. "If somebody says a complex word or asks you a question, like, 'What's 37 to the third power?' or something like that, then it'll pop up on the glasses." From the report: Ardayfio and Nguyen have raised $1 million to develop the glasses, led by Pillar VC, with support from Soma Capital, Village Global, and Morningside Venture. The glasses will be priced at $249 and will be available for preorder starting Wednesday. Ardayfio called the glasses "the first real step towards vibe thinking."

The two Ivy League dropouts, who have since moved into their own version of the Hacker Hostel in the San Francisco Bay Area, recently caused a stir after developing a facial-recognition app for Meta's smart Ray-Ban glasses to prove that the tech could be used to dox people. As a potential early competitor to Meta's smart glasses, Ardayfio said Meta, given its history of security and privacy scandals, had to rein in its product in ways that Halo can ultimately capitalize on. [...]

For now, Halo X glasses only have a display and a microphone, but no camera, although the two are exploring the possibility of adding it to a future model. Users still need to have their smartphones handy to help power the glasses and get "real time info prompts and answers to questions," per Nguyen. The glasses, which are manufactured by another company that the startup didn't name, are tethered to an accompanying app on the owner's phone, where the glasses essentially outsource the computing since they don't have enough power to do it on the device itself. Under the hood, the smart glasses use Google's Gemini and Perplexity as its chatbot engine, according to the two co-founders. Gemini is better for math and reasoning, whereas they use Perplexity to scrape the internet, they said.
Open Source

China's Lead in Open-Source AI Jolts Washington and Silicon Valley (msn.com) 89

China has established a lead in the field of open-source AI, a development that is reportedly sending jolts through both Washington and Silicon Valley. The nation's progress has become a significant event for American policymakers in the U.S. capital. The advancement has registered as a shock within Silicon Valley, the hub of the American technology industry. From the report: The overall performance of China's best open-weight model has surpassed the American open-source champion since November, according to research firm Artificial Analysis. The firm, which rates the ability of models in math, coding and other areas, found a version of Alibaba's Qwen3 beat OpenAI's gpt-oss.

However, the Chinese model is almost twice as big as OpenAI's, suggesting that for simpler tasks, Qwen might consume more computing power to do the same job. OpenAI said its open-source model outperformed rivals of similar size on reasoning tasks and delivered strong performance at low cost.

Microsoft

Microsoft Research Identifies 40 Jobs Most Vulnerable To AI (fortune.com) 166

Microsoft researchers have identified 40 occupations [PDF] with the highest exposure to AI, ranking jobs by how closely their tasks align with AI's current capabilities. The study analyzed 200,000 real-world conversations from Copilot users and compared AI performance against occupational data.

Interpreters and translators top the list, followed by historians and passenger attendants. Customer service and sales representatives, comprising about 5 million U.S. jobs, also face significant AI competition. Knowledge workers performing computer, math, or administrative tasks showed high vulnerability, as did sales positions involving information sharing and explanation. The research found occupations requiring Bachelor's degrees demonstrate higher AI applicability than those with lower educational requirements.

First, the top 10 least affected occupations by generative AI: 1. Dredge Operators
2. Bridge and Lock Tenders
3. Water Treatment Plant and System Operators
4. Foundry Mold and Coremakers
5. Rail-Track Laying and Maintenance Equipment Operators
6. Pile Driver Operators
7. Floor Sanders and Finishers
8. Orderlies
9. Motorboat Operators
10. Logging Equipment Operators
Now, the top 40 most affected occupations by generative AI:1. Interpreters and Translators
2. Historians
3. Passenger Attendants
4. Sales Representatives of Services
5. Writers and Authors
6. Customer Service Representatives
7. CNC Tool Programmers
8. Telephone Operators
9. Ticket Agents and Travel Clerks
10. Broadcast Announcers and Radio DJs
11. Brokerage Clerks
12. Farm and Home Management Educators
13. Telemarketers
14. Concierges
15. Political Scientists
16. News Analysts, Reporters, Journalists
17. Mathematicians
18. Technical Writers
19. Proofreaders and Copy Markers
20. Hosts and Hostesses
21. Editors
22. Business Teachers, Postsecondary
23. Public Relations Specialists
24. Demonstrators and Product Promoters
25. Advertising Sales Agents
26. New Accounts Clerks
27. Statistical Assistants
28. Counter and Rental Clerks
29. Data Scientists
30. Personal Financial Advisors
31. Archivists
32. Economics Teachers, Postsecondary
33. Web Developers
34. Management Analysts
35. Geographers
36. Models
37. Market Research Analysts
38. Public Safety Telecommunicators
39. Switchboard Operators
40. Library Science Teachers, Postsecondary.

Education

ChatGPT's New Study Mode Is Designed To Help You Learn, Not Just Give Answers 29

An anonymous reader quotes a report from Ars Technica: The rise of large language models like ChatGPT has led to widespread concern that "everyone is cheating their way through college," as a recent New York magazine article memorably put it. Now, OpenAI is rolling out a new "Study Mode" that it claims is less about providing answers or doing the work for students and more about helping them "build [a] deep understanding" of complex topics.

Study Mode isn't a new ChatGPT model but a series of "custom system instructions" written for the LLM "in collaboration with teachers, scientists, and pedagogy experts to reflect a core set of behaviors that support deeper learning," OpenAI said. Instead of the usual summary of a subject that stock ChatGPT might give -- which one OpenAI employee likened to "a mini textbook chapter" -- Study Mode slowly rolls out new information in a "scaffolded" structure. The mode is designed to ask "guiding questions" in the Socratic style and to pause for periodic "knowledge checks" and personalized feedback to make sure the user understands before moving on. It's unknown how many students will use this guided learning tool instead of just asking ChatGPT to generate answers from the start.

In an early hands-off demo attended by Ars Technica, Study Mode responded to a request to "teach me about game theory" by first asking about the user's overall familiarity with the subject and what they'll be using the information for. ChatGPT introduced a short overview of some core game theory concepts, then paused to ask a question before providing a relevant real-world example. In another example involving a classic "train traveling at speed" math problem, Study Mode resisted multiple simulated attempts by the frustrated "student" to simply ask for the answer and instead tried to gently redirect the conversation to how the available information could be used to generate that answer. An OpenAI representative told Ars that Study Mode will eventually provide direct solutions if asked repeatedly, but the default behavior is more tuned to a Socratic tutoring style.
OpenAI said it drew inspiration for Study Mode from "power users" and collaborated with pedagogy experts and college students to help refine its responses. As for whether the mode can be trusted, OpenAI told Ars that "the risk of hallucination is lower with Study Mode because the model processes information in smaller chunks, calibrating along the way."

The current Study Mode prompt does, however, result in some "inconsistent behavior and mistakes across conversations," the company warned.

Slashdot Top Deals