Vue normale

Il y a de nouveaux articles disponibles, cliquez pour rafraîchir la page.
Aujourd’hui — 3 juillet 2024Actualités numériques

British Startup Nyobolt Demos 4-Minute Battery Charging For EVs

Par : BeauHD
3 juillet 2024 à 00:30
Longtime Slashdot reader fahrbot-bot shares a report from CNN, written by Olesya Dmitracova: Nyobolt, based in Cambridge, has developed a new 35kWh lithium-ion battery that was charged from 10% to 80% in just over four and a half minutes in its first live demonstration last week. [...] Nyobolt's technology builds on a decade of research led by University of Cambridge battery scientist Clare Grey and Cambridge-educated Shivareddy, the company said. Key to its batteries' ability to be charged super-fast without a big impact on their longevity is a design that means they generate less heat. It also makes them safer as overheating can cause a lithium-ion battery to catch fire and explode. In addition, the materials used to make the batteries' anodes allow for a faster transfer of electrons. Nyobolt is currently in talks to sell its batteries to eight electric car manufacturers. At 35 kWh, the battery is much smaller than the 85 kWh in a more typical American electric vehicle (EV). Yet the technology may be used in larger battery packs in the future. Independent testing of Nyobolt's batteries by what it called a leading global manufacturer found that they can achieve over 4,000 fast-charge cycles, equivalent to 600,000 miles (965,600 kilometers), while retaining more than 80% of capacity, Nyobolt said in its Friday statement. William Kephart, an e-mobility specialist at consultancy P3 Group and a former engineer, said EV batteries of the kind Nyobolt has developed could "theoretically" be charged as fast as the firm is promising, but the challenge was manufacturing such batteries on an industrial scale. A crucial chemical element in Nyobolt's batteries is niobium but, as Kephart pointed out, last year only an estimated 83,000 tons (94,500 tons) was mined worldwide. Compare that with graphite, commonly used as anode material in lithium-ion batteries: an estimated 1.6 million tons (1.8 million tons) was produced in 2023. In addition, there are currently "a lot of unknowns" with the niobium battery technology, he told CNN. "The industry will work it out (but) it's not seen by the industry as a scalable technology just yet," he added.

Read more of this story at Slashdot.

Brazil Data Regulator Bans Meta From Mining Data To Train AI Models

Par : BeauHD
2 juillet 2024 à 23:50
Brazil's national data protection authority ruled on Tuesday that Meta must stop using data originating in the country to train its artificial intelligence models. The Associated Press reports: Meta's updated privacy policy enables the company to feed people's public posts into its AI systems. That practice will not be permitted in Brazil, however. The decision stems from "the imminent risk of serious and irreparable or difficult-to-repair damage to the fundamental rights of the affected data subjects," the agency said in the nation's official gazette. [...] Hye Jung Han, a Brazil-based researcher for the rights group, said in an email Tuesday that the regulator's action "helps to protect children from worrying that their personal data, shared with friends and family on Meta's platforms, might be used to inflict harm back on them in ways that are impossible to anticipate or guard against." But the decision regarding Meta will "very likely" encourage other companies to refrain from being transparent in the use of data in the future, said Ronaldo Lemos, of the Institute of Technology and Society of Rio de Janeiro, a think-tank. "Meta was severely punished for being the only one among the Big Tech companies to clearly and in advance notify in its privacy policy that it would use data from its platforms to train artificial intelligence," he said. Compliance must be demonstrated by the company within five working days from the notification of the decision, and the agency established a daily fine of 50,000 reais ($8,820) for failure to do so. In a statement, Meta said the company is "disappointed" by the decision and insists its method "complies with privacy laws and regulations in Brazil." "This is a step backwards for innovation, competition in AI development and further delays bringing the benefits of AI to people in Brazil," a spokesperson for the company added.

Read more of this story at Slashdot.

Phil Schiller To Join OpenAI Board In 'Observer' Role Following Apple's ChatGPT Deal

Par : BeauHD
2 juillet 2024 à 23:10
As reported by Bloomberg, Apple will get an "observer role" on OpenAI's board of directors as part of its partnership to integrate ChatGPT into iOS 18. That role will reportedly be filled by Apple Fellow, Phil Schiller. 9to5Mac reports: Apple having an "observer role" on the OpenAI board matches the role of Microsoft. Schiller will be able to observe and attend board meetings, but will not have any voting power: "The board observer role will put Apple on par with Microsoft, OpenAI's biggest backer and its main AI technology provider. The job allows someone to attend board meetings without being able to vote or exercise other director powers. Observers, however, do gain insights into how decisions are made at the company." The arrangement will take effect later this year, according to Bloomberg. Schiller "hasn't yet attended any meetings" of the OpenAI board and "details of the situation could still change." Schiller served as Apple's long-time marketing chief before transitioning to an Apple Fellow role in 2020. In this role, Schiller continues to lead the App Store and Apple events and reports directly to Apple CEO Tim Cook. Schiller is also leading Apple's efforts to defend the App Store against antitrust allegations around the world.

Read more of this story at Slashdot.

Google Emissions Jump Nearly 50% Over Five Years As AI Use Surges

Par : BeauHD
2 juillet 2024 à 22:30
An anonymous reader quotes a report from the Financial Times: Google's greenhouse gas emissions have surged 48 percent in the past five years due to the expansion of its data centers that underpin artificial intelligence systems, leaving its commitment to get to "net zero" by 2030 in doubt. The Silicon Valley company's pollution amounted to 14.3 million tons of carbon equivalent in 2023, a 48 percent increase from its 2019 baseline and a 13 percent rise since last year, Google said in its annual environmental report on Tuesday. Google said the jump highlighted "the challenge of reducing emissions" at the same time as it invests in the build-out of large language models and their associated applications and infrastructure, admitting that "the future environmental impact of AI" was "complex and difficult to predict." Chief sustainability officer Kate Brandt said the company remained committed to the 2030 target but stressed the "extremely ambitious" nature of the goal. "We do still expect our emissions to continue to rise before dropping towards our goal," said Brandt. She added that Google was "working very hard" on reducing its emissions, including by signing deals for clean energy. There was also a "tremendous opportunity for climate solutions that are enabled by AI," said Brandt. [...] In Tuesday's report, Google said its 2023 energy-related emissions -- which come primarily from data center electricity consumption -- rose 37 percent year on year, and overall represented a quarter of its total greenhouse gas emissions. Google's supply chain emissions -- its largest chunk, representing 75 percent of its total emissions -- also rose 8 percent. Google said they would "continue to rise in the near term" as a result in part of the build-out of the infrastructure needed to run AI systems. Google has pledged to achieve net zero across its direct and indirect greenhouse gas emissions by 2030, and to run on carbon-free energy during every hour of every day within each grid it operates by the same date. However, the company warned in Tuesday's report that the "termination" of some clean energy projects during 2023 had pushed down the amount of renewables it had access to. Meanwhile, the company's data centre electricity consumption had "outpaced" Google's ability to bring more clean power projects online in the US and Asia-Pacific regions. Google's data centre electricity consumption increased 17 percent in 2023, and amounted to approximately 7-10 percent of global data center electricity consumption, the company estimated.Its data centers also consumed 17 percent more water in 2023 than during the previous year, Google said.

Read more of this story at Slashdot.

Hier — 2 juillet 2024Actualités numériques

Bruce Bastian, WordPerfect Co-Creator, Dies At 76

Par : BeauHD
2 juillet 2024 à 21:50
Longtime Slashdot reader regoli shares an obituary from the Wall Street Journal: When Alan Ashton was a computer-science professor at Brigham Young University in the mid-1970s, the director of the school's marching band knocked on his door and said he wanted to use a computer to choreograph the band's halftime shows. Ashton was easily persuaded; he was a trumpet player whose Ph.D. dissertation at the University of Utah was "Electronics, music and computers." Bruce Bastian, the graduate student who was working as BYU's marching-band director, turned out to be a quick learner. "He was very conscientious, very thorough," Ashton said in an interview, "and just absolutely brilliant." Within a few years, the two were at work on a program that would turn them into two of the richest people in the nation, founders of the company that made WordPerfect, the dominant word-processing software in the 1980s and early '90s and one of the first pieces of software many Americans bought when they brought home their first PCs. But behind the hundreds of millions of dollars and blockbuster success, Bastian's personal life, he later said, was in "free fall." Between the time he and Ashton released what would later be known as WordPerfect to the public in 1980 and when they sold the company for $1.4 billion in 1994, Bastian told his wife, four sons and his Mormon community that he was gay and leaving both his marriage and the church. When he died, June 16, at the age of 76 from complications associated with pulmonary fibrosis, he had been living a different life: A longtime advocate for LGBTQ rights, Bastian was married to a man, and had found a way to maintain relationships with his family, who remained dedicated members of the church that rejected his sexual orientation. "I kind of have three parts of my life," he said in 2010 during one of several extensive interviews he gave to the Mormon Stories podcast, "the pre-WordPerfect life, the WordPerfect years, and now the LGBT years." Other publications remembering Bruce Bastian include: The New York Times, The Salt Lake Tribune, University of Utah, and Human Rights Campaign.

Read more of this story at Slashdot.

Supreme Court to Hear Case on Texas Law Restricting Access to Porn

Par : BeauHD
2 juillet 2024 à 21:10
The Supreme Court has agreed to hear a challenge to a Texas law requiring age verification to access online pornography, which opponents argue violates the First Amendment by discouraging adults from viewing such material due to privacy concerns. A federal judge blocked the law citing its chilling effect on free speech, but a divided appeals court upheld it, emphasizing the government's interest in protecting minors; the case will now be reviewed by the Supreme Court. The Texas bill in question, HB 1181, was passed into law last June. The New York Times reports: The Supreme Court agreed on Tuesday to hear a challenge to a Texas law that seeks to limit minors' access to pornography on the internet by requiring age verification measures like the submission of government-issued IDs. A trade group, companies that produce sexual materials and a performer challenged the law, saying that it violates the First Amendment right of adults. The law does not allow companies to retain information their users submit. But the challengers said adults would be wary of supplying personal information for fear of identity theft, tracking and extortion. [...] In urging the Supreme Court to leave the law in place while it considers whether to hear the case, Ken Paxton, Texas' attorney general, said pornography available on the internet is "orders of magnitude more graphic, violent and degrading than any so-called 'girlie' magazine of yesteryear." He added: "This statute does not prohibit the performance, production or even sale of pornography but, more modestly, simply requires the pornography industry that make billions of dollars from peddling smut to take commercially reasonable steps to ensure that those who access the material are adults. There is nothing unconstitutional about it."

Read more of this story at Slashdot.

AI Trains On Kids' Photos Even When Parents Use Strict Privacy Settings

Par : BeauHD
2 juillet 2024 à 20:33
An anonymous reader quotes a report from Ars Technica: Human Rights Watch (HRW) continues to reveal how photos of real children casually posted online years ago are being used to train AI models powering image generators -- even when platforms prohibit scraping and families use strict privacy settings. Last month, HRW researcher Hye Jung Han found 170 photos of Brazilian kids that were linked in LAION-5B, a popular AI dataset built from Common Crawl snapshots of the public web. Now, she has released a second report, flagging 190 photos of children from all of Australia's states and territories, including indigenous children who may be particularly vulnerable to harms. These photos are linked in the dataset "without the knowledge or consent of the children or their families." They span the entirety of childhood, making it possible for AI image generators to generate realistic deepfakes of real Australian children, Han's report said. Perhaps even more concerning, the URLs in the dataset sometimes reveal identifying information about children, including their names and locations where photos were shot, making it easy to track down children whose images might not otherwise be discoverable online. That puts children in danger of privacy and safety risks, Han said, and some parents thinking they've protected their kids' privacy online may not realize that these risks exist. From a single link to one photo that showed "two boys, ages 3 and 4, grinning from ear to ear as they hold paintbrushes in front of a colorful mural," Han could trace "both children's full names and ages, and the name of the preschool they attend in Perth, in Western Australia." And perhaps most disturbingly, "information about these children does not appear to exist anywhere else on the Internet" -- suggesting that families were particularly cautious in shielding these boys' identities online. Stricter privacy settings were used in another image that Han found linked in the dataset. The photo showed "a close-up of two boys making funny faces, captured from a video posted on YouTube of teenagers celebrating" during the week after their final exams, Han reported. Whoever posted that YouTube video adjusted privacy settings so that it would be "unlisted" and would not appear in searches. Only someone with a link to the video was supposed to have access, but that didn't stop Common Crawl from archiving the image, nor did YouTube policies prohibiting AI scraping or harvesting of identifying information. Reached for comment, YouTube's spokesperson, Jack Malon, told Ars that YouTube has "been clear that the unauthorized scraping of YouTube content is a violation of our Terms of Service, and we continue to take action against this type of abuse." But Han worries that even if YouTube did join efforts to remove images of children from the dataset, the damage has been done, since AI tools have already trained on them. That's why -- even more than parents need tech companies to up their game blocking AI training -- kids need regulators to intervene and stop training before it happens, Han's report said. Han's report comes a month before Australia is expected to release a reformed draft of the country's Privacy Act. Those reforms include a draft of Australia's first child data protection law, known as the Children's Online Privacy Code, but Han told Ars that even people involved in long-running discussions about reforms aren't "actually sure how much the government is going to announce in August." "Children in Australia are waiting with bated breath to see if the government will adopt protections for them," Han said, emphasizing in her report that "children should not have to live in fear that their photos might be stolen and weaponized against them."

Read more of this story at Slashdot.

10-Year-Old Open Source Flaw Could Affect 'Almost Every Apple Device'

Par : BeauHD
2 juillet 2024 à 13:00
storagedude shares a report from the Cyber Express: Some of the most widely used web and social media applications could be vulnerable to three newly discovered CocoaPods vulnerabilities -- including potentially millions of Apple devices, according to a report by The Cyber Express, the news service of threat intelligence vendor Cyble Inc. E.V.A Information Security researchers reported three vulnerabilities in the open source CocoaPods dependency manager that could allow malicious actors to take over thousands of unclaimed pods and insert malicious code into many of the most popular iOS and MacOS applications, potentially affecting "almost every Apple device." The researchers found vulnerable code in applications provided by Meta (Facebook, Whatsapp), Apple (Safari, AppleTV, Xcode), and Microsoft (Teams); as well as in TikTok, Snapchat, Amazon, LinkedIn, Netflix, Okta, Yahoo, Zynga, and many more. The vulnerabilities have been patched, yet the researchers still found 685 Pods "that had an explicit dependency using an orphaned Pod; doubtless there are hundreds or thousands more in proprietary codebases." The newly discovered vulnerabilities -- one of which (CVE-2024-38366) received a 10 out of 10 criticality score -- actually date from a May 2014 CocoaPods migration to a new 'Trunk' server, which left 1,866 orphaned pods that owners never reclaimed. While the vulnerabilities have been patched, the work for developers and DevOps teams that used CocoaPods before October 2023 is just getting started. "Developers and DevOps teams that have used CocoaPods in recent years should verify the integrity of open source dependencies used in their application code," the E.V.A researchers said. "The vulnerabilities we discovered could be used to control the dependency manager itself, and any published package." [...] "Dependency managers are an often-overlooked aspect of software supply chain security," the researchers wrote. "Security leaders should explore ways to increase governance and oversight over the use these tools." "While there is no direct evidence of any of these vulnerabilities being exploited in the wild, evidence of absence is not absence of evidence." the EVA researchers wrote. "Potential code changes could affect millions of Apple devices around the world across iPhone, Mac, AppleTV, and AppleWatch devices." While no action is required by app developers or users, the EVA researchers recommend several ways to protect against these vulnerabilities. To ensure secure and consistent use of CocoaPods, synchronize the podfile.lock file with all developers, perform CRC validation for internally developed Pods, and conduct thorough security reviews of third-party code and dependencies. Furthermore, regularly review and verify the maintenance status and ownership of CocoaPods dependencies, perform periodic security scans, and be cautious of widely used dependencies as potential attack targets.

Read more of this story at Slashdot.

Swiss Broadcasting Corporation To Pull Plug On FM Radio

Par : BeauHD
2 juillet 2024 à 10:00
Digital audio broadcasting (DAB+) and internet radio have largely replaced traditional FM radio in Switzerland, with digital radio holding an 81% share of use in spring 2023. Due to the high costs of maintaining FM transmitters and declining financial resources, Switzerland plans to fully transition to digital radio by the end of 2026, phasing out FM broadcasting completely. From a report: DAB+ and the internet offer better quality and a larger program selection, are more energy and cost efficient, and can provide additional information in text and images, it said. To receive DAB+ requires a corresponding device or adapter, and new cars have been equipped with digital technology as standard for several years. In addition, the Federal Roads Office (FEDRO) will upgrade all tunnels on the national road network for digital radio reception by the end of the year and switch off FM transmitters. FM was originally expected to be switched off throughout Switzerland by the end of 2024. The government extended FM licenses for the radio industry for the last time in October 2023 to the end of 2026, after which radio stations in Switzerland will no longer be able to broadcast via FM, only digitally. OFCOM announced at the time that the final extension would give the radio industry the flexibility to complete the transition process from analogue to digital radio.

Read more of this story at Slashdot.

Arctic 'Dirty Fuel' Ban For Ships Comes Into Force

Par : BeauHD
2 juillet 2024 à 07:00
Starting July 1st, ships in Arctic waters are banned from using Heavy Fuel Oil (HFO), a relatively cheap tar-like oil that's widely used in shipping around the world, especially tankers. According to the BBC, it's the "dirtiest and most climate-damaging fuel for ships." Still, campaigners believe numerous loopholes will allow most ships to continue using the fuel until 2029, limiting the ban's immediate effectiveness. The BBC reports: Produced from the waste left over in oil refining, HFO poses a huge threat to the oceans in general but to the Arctic in particular. This sludge-like fuel is almost impossible to clean up if a spill occurs. In colder waters, experts say, the fuel does not break down but sinks in lumps that linger in sediments, threatening fragile ecosystems. In climate terms, this oil is seen as particularly dangerous, not just producing large amounts of planet-warming gas when burned, but also spewing out sooty particles called black carbon. [...] The oil was banned from use or transport in the Antarctic in 2011. Environmentalists have been pushing to expand that restriction to northern waters for years, finally persuading the countries that participate in the International Maritime Organisation (IMO) to enact a ban back in 2021. [...] According to the regulations, ships that have a "protected fuel tank" will be exempt from the ban. Countries that border the Arctic will also be able to exempt their own ships from the ban in their own territorial waters. One of the major players in the region is Russia, which has over 800 ships operating in northern waters. They are not implementing the new IMO regulation. These waiver exemptions will last until 2029 -- their impact is likely to be significant, with the International Council on Clean Transportation estimating that about 74% of ships that use HFO will be able to continue to do so. Some observers believe that increased efforts to extract oil in the Arctic could see a rise in the amount of HFO in use in these waters, instead of a decrease.

Read more of this story at Slashdot.

Survey Finds Public Perception of Scientists' Credibility Has Slipped

Par : BeauHD
2 juillet 2024 à 03:30
An anonymous reader quotes a report from Phys.Org: New analyses from the Annenberg Public Policy Center find that public perceptions of scientists' credibility -- measured as their competence, trustworthiness, and the extent to which they are perceived to share an individual's values -- remain high, but their perceived competence and trustworthiness eroded somewhat between 2023 and 2024. The research also found that public perceptions of scientists working in artificial intelligence (AI) differ from those of scientists as a whole. [...] The five factors in Factors Assessing Science's Self-Presentation (FASS) are whether science and scientists are perceived to be credible and prudent, and whether they are perceived to overcome bias, correct error (self-correcting), and whether their work benefits people like the respondent and the country as a whole (beneficial). [...] In the FASS model, perceptions of scientists' credibility are assessed through perceptions of whether scientists are competent, trustworthy, and "share my values." The first two of those values slipped in the most recent survey. In 2024, 70% of those surveyed strongly or somewhat agree that scientists are competent (down from 77% in 2023) and 59% strongly or somewhat agree that scientists are trustworthy (down from 67% in 2023). The survey also found that in 2024, fewer people felt that scientists' findings benefit "the country as a whole" and "benefit people like me." In 2024, 66% strongly or somewhat agreed that findings benefit the country as a whole (down from 75% in 2023). Belief that scientists' findings "benefit people like me," also declined, to 60% from 68%. Taken together, those two questions make up the beneficial factor of FASS. The findings follow sustained attacks on climate and COVID-19-related science, and more recently, public concerns about the rapid development and deployment of artificial intelligence. Here's what the study found when comparing perceptions of scientists in general with climate and AI scientists: - Credibility: When asked about three factors underlying scientists' credibility, AI scientists have lower credibility in all three values. - Competent: 0% strongly/somewhat agree that scientists are competent, but only 62% for climate scientists and 49% for AI scientists. - Trustworthy: 59% agree scientists are trustworthy, 54% agree for climate scientists, 28% for AI scientists. - Share my values: A higher number (38%) agree that climate scientists share my values than for scientists in general (36%) and AI scientists (15%). More people disagree with this for AI scientists (35%) than for the others. - Prudence: Asked whether they agree or disagree that science by various groups of scientists "creates unintended consequences and replaces older problems with new ones," over half of those surveyed (59%) agree that AI scientists create unintended consequences and just 9% disagree. - Overcoming bias: Just 42% of those surveyed agree that scientists "are able to overcome human and political biases," but only 21% feel that way about AI scientists. In fact, 41% disagree that AI scientists are able to overcome human political biases. In another area, just 23% agree that AI scientists provide unbiased conclusions in their area of inquiry and 38% disagree. - Self-correction: Self-correction, or "organized skepticism expressed in expectations sustaining a culture of critique," as the FASS paper puts it, is considered by some as a "hallmark of science." AI scientists are seen as less likely than scientists or climate scientists to take action to prevent fraud; take responsibility for mistakes; or to have mistakes that are caught by peer review. - Benefits: Asked about the benefits from scientists' findings, 60% agree that scientists' "findings benefit people like me," though just 44% agree for climate scientists and 35% for AI scientists. Asked about whether findings benefit the country as a whole, 66% agree for scientists, 50% for climate scientists and 41% for AI scientists. - Your best interest: The survey also asked respondents how much trust they have in scientists to act in the best interest of people like you. (This specific trust measure is not a part of the FASS battery.) Respondents have less trust in AI scientists than in others: 41% have a great deal/a lot of trust in medical scientists; 39% in climate scientists; 36% in scientists; and 12% in AI scientists.

Read more of this story at Slashdot.

Alzheimer's Scientist Indicted For Allegedly Falsifying Data In $16 Million Scheme

Par : BeauHD
1 juillet 2024 à 23:20
"A federal grand jury has indicted an embattled Alzheimer's researcher for allegedly falsifying data to fraudulently obtain $16 million in federal research funding from the National Institutes of Health for the development of a controversial Alzheimer's drug and diagnostic test," writes Beth Mole via Ars Technica. "Wang is charged with one count of major fraud against the United States, two counts of wire fraud, and one count of false statements. If convicted, he faces a maximum penalty of 10 years in prison for the major fraud charge, 20 years in prison for each count of wire fraud, and five years in prison for the count of false statements [...]." From the report: Hoau-Yan Wang, 67, a medical professor at the City University of New York, was a paid collaborator with the Austin, Texas-based pharmaceutical company Cassava Sciences. Wang's research and publications provided scientific underpinnings for Cassava's Alzheimer's treatment, Simufilam, which is now in Phase III trials. Simufilam is a small-molecule drug that Cassava claims can restore the structure and function of a scaffolding protein in the brain of people with Alzheimer's, leading to slowed cognitive decline. But outside researchers have long expressed doubts and concerns about the research. In 2023, Science magazine obtained a 50-page report from an internal investigation at CUNY that looked into 31 misconduct allegations made against Wang in 2021. According to the report, the investigating committee "found evidence highly suggestive of deliberate scientific misconduct by Wang for 14 of the 31 allegations," the report states. The allegations largely centered around doctored and fabricated images from Western blotting, an analytical technique used to separate and detect proteins. However, the committee couldn't conclusively prove the images were falsified "due to the failure of Dr. Wang to provide underlying, original data or research records and the low quality of the published images that had to be examined in their place." In all, the investigation "revealed long-standing and egregious misconduct in data management and record keeping by Dr. Wang," and concluded that "the integrity of Dr. Wang's work remains highly questionable." The committee also concluded that Cassava's lead scientist on its Alzheimer's disease program, Lindsay Burns, who was a frequent co-author with Wang, also likely bears some responsibility for the misconduct. In March 2022, five of Wang's articles published in the journal PLOS One were retracted over integrity concerns with images in the papers. Other papers by Wang have also been retracted or had statements of concern attached to them. Further, in September 2022, the Food and Drug Administration conducted an inspection of the analytical work and techniques used by Wang to analyze blood and cerebrospinal fluid from patients in a simufilam trial. The investigation found a slew of egregious problems, which were laid out in a "damning" report (PDF) obtained by Science. In the indictment last week (PDF), federal authorities were explicit about the allegations, claiming that Wang falsified the results of his scientific research to NIH "by, among other things, manipulating data and images of Western blots to artificially add bands [which represent proteins], subtract bands, and change their relative thickness and/or darkness, and then drawing conclusions" based on those false results.

Read more of this story at Slashdot.

À partir d’avant-hierActualités numériques

Microsoft Tells Yet More Customers Their Emails Have Been Stolen

Par : BeauHD
1 juillet 2024 à 22:40
Microsoft revealed that the Russian hackers who breached its systems earlier this year stole more emails than initially reported. "We are continuing notifications to customers who corresponded with Microsoft corporate email accounts that were exfiltrated by the Midnight Blizzard threat actor, and we are providing the customers the email correspondence that was accessed by this actor," a Microsoft spokesperson told Bloomberg (paywalled). "This is increased detail for customers who have already been notified and also includes new notifications." The Register reports: We've been aware for some time that the digital Russian break-in at the Windows maker saw Kremlin spies make off with source code, executive emails, and sensitive U.S. government data. Reports last week revealed that the issue was even larger than initially believed and additional customers' data has been stolen. Along with Russia, Microsoft was also compromised by state actors from China not long ago, and that issue similarly led to the theft of emails and other data belonging to senior U.S. government officials. Both incidents have led experts to call Microsoft a threat to U.S. national security, and president Brad Smith to issue a less-than-reassuring mea culpa to Congress. All the while, the U.S. government has actually invested more in its Microsoft kit. Bloomberg reported that emails being sent to affected Microsoft customers include a link to a secure environment where customers can visit a site to review messages Microsoft identified as having been compromised. But even that might not have been the most security-conscious way to notify folks: Several thought they were being phished.

Read more of this story at Slashdot.

'Roaring Kitty' Is Sued For Alleged GameStop Manipulation

Par : BeauHD
1 juillet 2024 à 22:00
Keith Gill, the investor known as "Roaring Kitty" online, is being used by GameStop investors for helping spur the meme stock mania of 2021. The plaintiffs said they lost money through his "pump-and-dump" scheme, which led to a "short squeeze" that caused losses for hedge funds betting stock prices would fall. Reuters reports: A proposed class action accusing Gill of securities fraud was filed on Friday in the Brooklyn, New York federal court. Investors led by Martin Radev, who lives in the Las Vegas area, said Gill manipulated GameStop securities between May 13 and June 13 by quietly accumulating large quantities of stock and call options, and then dumping some holdings after emerging from a three-year social media hiatus. They said Gill's activities caused GameStop's share price to gyrate wildly, generating "millions of dollars" in profit for him at their expense. "Defendant still enjoys celebrity status and commands a following of millions through his social media accounts," the complaint said. "Accordingly, Defendant was well aware of his ability to manipulate the market for GameStop securities, as well as the benefits he could reap." He had on May 12 posted a cryptic meme on the social media platform X that was widely seen as a bullish signal for GameStop, whose stock he cheerleaded in 2021. GameStop's share price more than tripled over the next two days, but gave back nearly all the gains by May 24. On June 2, Gill revealed that he owned 5 million GameStop shares and 120,000 call options, and on June 13 revealed he had shed the call options but owned 9 million GameStop shares. Investors said the truth about Gill's investing became known on June 3 when the Wall Street Journal wrote about the timing of his options trades and said the online brokerage E*Trade considered kicking him off its platform.

Read more of this story at Slashdot.

'Julian Assange Should Not Have Been Prosecuted In the First Place'

Par : BeauHD
1 juillet 2024 à 21:20
An anonymous reader quotes an op-ed written by Kenneth Roth, former executive director of Human Rights Watch (1993-2022) and a visiting professor at Princeton's School of Public and International Affairs: Julian Assange's lengthy detention has finally ended, but the danger that his prosecution poses to the rights of journalists remains. As is widely known, the U.S. government's pursuit of Assange under the Espionage Act threatens to criminalize common journalistic practices. Sadly, Assange's guilty plea and release from custody have done nothing to ease that threat. That Assange was indicted under the Espionage Act, a U.S. law designed to punish spies and traitors, should not be considered the normal course of business. Barack Obama's justice department never charged Assange because it couldn't distinguish what he had done from ordinary journalism. The espionage charges were filed by the justice department of Donald Trump. Joe Biden could have reverted to the Obama position and withdrawn the charges but never did. The 18-count indictment filed under Trump accused Assange of having solicited secret U.S. government information and encouraged Chelsea Manning to provide it. Manning committed a crime when she delivered that information because she was a government employee who had pledged to safeguard confidential information on pain of punishment. But Assange's alleged solicitation of that information, and the steps he was said to have taken to ensure that it could be transferred anonymously, are common procedure for many journalists who report on national security issues. If these practices were to be criminalized, our ability to monitor government conduct would be seriously compromised. To make matters worse, someone accused under the Espionage Act is not allowed to argue to a jury that disclosures were made in the public interest. The unauthorized disclosure of secret information deemed prejudicial to national security is sufficient for conviction regardless of motive. To justify Espionage Act charges, the Trump-era prosecutors stressed that Assange was accused of not only soliciting and receiving secret government information but also agreeing to help crack a password that would provide access to U.S. government files. That is not ordinary journalistic behavior. An Espionage Act prosecution for computer hacking is very different from a prosecution for merely soliciting and receiving secret information. Even if it would not withdraw the Trump-era charges, Biden's justice department could have limited the harm to journalistic freedom by ensuring that the alleged computer hacking was at the center of Assange's guilty plea. In fact, it was nowhere to be found. The terms for the proceeding were outlined in a 23-page "plea agreement" filed with the U.S. District Court for the Northern Mariana Islands, where Assange appeared by consent. Assange agreed to plead guilty to a single charge of violating the Espionage Act, but under U.S. law, it is not enough to plead in the abstract. A suspect must concede facts that would constitute an offense. "One effect of the guilty plea is that there will be no legal challenge to the prosecution, and hence no judicial decision on whether this use of the Espionage Act violates the freedom of the media as protected by the first amendment of the U.S. constitution," notes Roth. "That means that just as prosecutors overreached in the case of Assange, they could do so again." "[M]edia protections are not limited to journalists who are deemed responsible. Nor do we want governments to make judgments about which journalists deserve First Amendment safeguards. That would quickly compromise media freedom for all journalists." Roth concludes: "Imperfect journalist that he was, Assange should never have been prosecuted under the Espionage Act. It is unfortunate that the Biden administration didn't take available steps to mitigate that harm."

Read more of this story at Slashdot.

The Vision Pro Will Get Apple Intelligence, 'Go Deeper' In-Store Demos

Par : BeauHD
1 juillet 2024 à 20:40
According to Bloomberg's Mark Gurman, Apple plans to add its "Apple Intelligence" AI features to visionOS and update its approach to in-store demos of the headset. The Verge reports: The company is adding a new "Go Deeper" option to its in-store demos, Gurman writes. That reportedly includes testing office features and watching videos, as well as defaulting to the Dual Loop band that sends straps over the top and around the back of wearers' heads instead of the single-strap Solo Loop band, which some find uncomfortable. Apple will also reportedly let people view their own videos and photos, including panoramas, in the headset. Adding the sentimental touch to the demos could work out, especially once visionOS 2 comes out this fall, with its "spatialize" option to turn 2D photos into 3D ones -- a feature that's more impressive than it has the right to be (though still a little quirky with hair and glasses, like Apple's Portrait Mode feature).

Read more of this story at Slashdot.

Redbox Owner Chicken Soup For the Soul Files For Chapter 11 Bankruptcy Protection

Par : BeauHD
1 juillet 2024 à 20:00
Chicken Soup for the Soul Entertainment, the parent of DVD rental operator Redbox, has filed for Chapter 11 bankruptcy protection after months of financial struggles and piling unpaid bills. The Associated Press reports: Chicken Soup for the Soul has accumulated nearly $1 billion in debt, the Chapter 11 filing submitted Friday in Delaware bankruptcy court shows, after reporting loss after loss over recent quarters. The filing also discloses that Chicken Soup for the Soul owes millions to over 500 creditors -- which range from big names in the entertainment world like Sony Pictures and Warner Bros, to major retailers like Walgreens and Walmart. As of March of this year, Friday's filing shows, Chicken Soup for the Soul had about $414 million in assets and $970 million in debts. Shares for the public company have fallen more than 90% over the last year. "Redbox, founded in 2002, is best known for red-colored, self-serve machines that sit outside of pharmacies or groceries stores to rent or sell DVDs," notes the report. It was acquired by Chicken Soup for the Soul in 2022. There are currently about 27,000 Redbox kiosks across the U.S. -- down from 36,000 at the Redbox acquisition was finalized in August 2022.

Read more of this story at Slashdot.

Meta Defends Charging Fee For Privacy Amid Showdown With EU

Par : BeauHD
1 juillet 2024 à 19:18
An anonymous reader quotes a report from Ars Technica: Meta continues to hit walls with its heavily scrutinized plan to comply with the European Union's strict online competition law, the Digital Markets Act (DMA), by offering Facebook and Instagram subscriptions as an alternative for privacy-inclined users who want to opt out of ad targeting. Today, the European Commission (EC) announced preliminary findings that Meta's so-called "pay or consent" or "pay or OK" model -- which gives users a choice to either pay for access to its platforms or give consent to collect user data to target ads -- is not compliant with the DMA. According to the EC, Meta's advertising model violates the DMA in two ways. First, it "does not allow users to opt for a service that uses less of their personal data but is otherwise equivalent to the 'personalized ads-based service." And second, it "does not allow users to exercise their right to freely consent to the combination of their personal data," the press release said. Now, Meta will have a chance to review the EC's evidence and defend its policy, with today's findings kicking off a process that will take months. The EC's investigation is expected to conclude next March. Thierry Breton, the commissioner for the internal market, said in the press release that the preliminary findings represent "another important step" to ensure Meta's full compliance with the DMA. "The DMA is there to give back to the users the power to decide how their data is used and ensure innovative companies can compete on equal footing with tech giants on data access," Breton said. A Meta spokesperson told Ars that Meta plans to fight the findings -- which could trigger fines up to 10 percent of the company's worldwide turnover, as well as fines up to 20 percent for repeat infringement if Meta loses. The EC agreed that more talks were needed, writing in the press release, "the Commission continues its constructive engagement with Meta to identify a satisfactory path towards effective compliance." Meta continues to claim that its "subscription for no ads" model was "endorsed" by the highest court in Europe, the Court of Justice of the European Union (CJEU), last year. "Subscription for no ads follows the direction of the highest court in Europe and complies with the DMA," Meta's spokesperson said. "We look forward to further constructive dialogue with the European Commission to bring this investigation to a close." Meta rolled out its ad-free subscription service option last November. "Depending on where you purchase it will cost $10.5/month on the web or $13.75/month on iOS and Android," said the company in a blog post. "Regardless of where you purchase, the subscription will apply to all linked Facebook and Instagram accounts in a user's Accounts Center. As is the case for many online subscriptions, the iOS and Android pricing take into account the fees that Apple and Google charge through respective purchasing policies."

Read more of this story at Slashdot.

ChatGPT Outperforms Undergrads In Intro-Level Courses, Falls Short Later

Par : BeauHD
29 juin 2024 à 10:00
Peter Scarfe, a researcher at the University of Reading's School of Psychology and Clinical Language Sciences, conducted an experiment testing the vulnerability of their examination system to AI-generated work. Using ChatGPT-4, Scarfe's team submitted over 30 AI-generated answers across multiple undergraduate psychology modules, finding that 94 percent of these submissions went undetected and nearly 84 percent received higher grades than human counterparts. The findings have been published in the journal PLOS One. Ars Technica reports: Scarfe's team submitted AI-generated work in five undergraduate modules, covering classes needed during all three years of study for a bachelor's degree in psychology. The assignments were either 200-word answers to short questions or more elaborate essays, roughly 1,500 words long. "The markers of the exams didn't know about the experiment. In a way, participants in the study didn't know they were participating in the study, but we've got necessary permissions to go ahead with that," Scarfe claims. Shorter submissions were prepared simply by copy-pasting the examination questions into ChatGPT-4 along with a prompt to keep the answer under 160 words. The essays were solicited the same way, but the required word count was increased to 2,000. Setting the limits this way, Scarfe's team could get ChatGPT-4 to produce content close enough to the required length. "The idea was to submit those answers without any editing at all, apart from the essays, where we applied minimal formatting," says Scarfe. Overall, Scarfe and his colleagues slipped 63 AI-generated submissions into the examination system. Even with no editing or efforts to hide the AI usage, 94 percent of those went undetected, and nearly 84 percent got better grades (roughly half a grade better) than a randomly selected group of students who took the same exam. "We did a series of debriefing meetings with people marking those exams and they were quite surprised," says Scarfe. Part of the reason they were surprised was that most of those AI submissions that were detected did not end up flagged because they were too repetitive or robotic -- they got flagged because they were too good. Out of five modules where Scarfe's team submitted AI work, there was one where it did not receive better grades than human students: the final module taken by students just before they left the university. "Large language models can emulate human critical thinking, analysis, and integration of knowledge drawn from different sources to a limited extent. In their last year at the university, students are expected to provide deeper insights and use more elaborate analytical skills. The AI isn't very good at that, which is why students fared better," Scarfe explained. All those good grades Chat GPT-4 got were in the first- and second-year exams, where the questions were easier. "But the AI is constantly improving, so it's likely going to score better in those advanced assignments in the future. And since AI is becoming part of our lives and we don't really have the means to detect AI cheating, at some point we are going to have to integrate it into our education system," argues Scarfe. He said the role of a modern university is to prepare the students for their professional careers, and the reality is they are going to use various AI tools after graduation. So, they'd be better off knowing how to do it properly.

Read more of this story at Slashdot.

❌
❌