利用人工智慧製造假綁架案
本週,在一場(chǎng)離奇的網(wǎng)路綁架事件中,一名失蹤的 17 歲中國交換生在猶他州冰凍荒野的帳篷裡被發(fā)現(xiàn)還活著。 他被騙子操縱進(jìn)入荒野,騙子聲稱綁架了他,向他的父母勒索了 8 萬美元的贖金。
Riverdale 警方正在處理網(wǎng)路綁架現(xiàn)象。
雖然尚不清楚這起事件中是否使用了人工智慧,但它揭示了假綁架事件不斷增加的趨勢(shì),這些綁架事件通常針對(duì)中國交換生。 Riverdale 警察局表示,詐騙者經(jīng)常透過威脅傷害家人來說服受害者孤立自己,然後使用恐懼策略和偽造的照片和音訊(有時(shí)是上演的,有時(shí)是由被綁架受害者的人工智慧產(chǎn)生的)來勒索金錢。亞利桑那州婦女 Jennifer DeStefano 去年,她向美國參議院作證稱,她被 Deepfake 人工智慧技術(shù)愚弄,以為自己 15 歲的女兒 Briana 被綁架了。 詐騙者顯然得知這名青少年正在滑雪旅行,然後用模仿布萊恩娜抽泣和哭泣的深度偽造人工智能聲音給詹妮弗打電話:媽媽,這些壞人抓住了我,幫助我,幫助我。
隨後,一名男子威脅說,如果不支付贖金,就要給布里安娜注射大量毒品並殺死她。 幸運(yùn)的是,在她交出任何現(xiàn)金之前,另一位家長(zhǎng)提到他們聽說過類似的人工智慧騙局,珍妮佛能夠聯(lián)繫到真正的布萊恩娜,得知她是安全的。 警方對(duì)她的報(bào)告不感興趣,稱其為惡作劇電話。
劍橋大學(xué)心理學(xué)教授桑德范德林登建議人們避免在網(wǎng)路上發(fā)布旅行計(jì)劃,並盡可能少地向來電者發(fā)送垃圾郵件,以阻止他們捕捉你的聲音。 如果您在網(wǎng)路上有大量音訊或視訊片段,您可能需要考慮將其刪除。
機(jī)器人技術(shù)的「ChatGPT 時(shí)刻」?
Figure Robot 的創(chuàng)始人 Brett Adcock 週末用小寫字母氣喘吁籲地發(fā)推文說,我們的實(shí)驗(yàn)室剛剛?cè)〉昧巳斯ぶ悄芡黄疲瑱C(jī)器人技術(shù)即將迎來它的 ChatGPT 時(shí)刻。這可能有點(diǎn)過分誇大了。 這項(xiàng)突破在一段一分鐘的影片中得以體現(xiàn),影片展示了該公司的Figure-01機(jī)器人在接受人類10小時(shí)的指導(dǎo)後,自行製作咖啡。
製作咖啡並不是那麼具有開創(chuàng)性(當(dāng)然也不是每個(gè)人都印象深刻),但影片聲稱機(jī)器人能夠從錯(cuò)誤中學(xué)習(xí)並自我糾正。 因此,當(dāng)圖 01 將咖啡包放錯(cuò)時(shí),它會(huì)足夠聰明地輕推它,使其進(jìn)入插槽。 到目前為止,人工智慧在自我糾正錯(cuò)誤方面表現(xiàn)相當(dāng)糟糕。
Adcock said: The reason why this is so groundbreaking is if you can get human data for an application (making coffee, folding laundry, warehouse work, etc), you can then train an AI system end-to-end, on Figure-01 there is a path to scale to every use case and when the fleet expands, further data is collected from the robot fleet, re-trained, and the robot achieves even better performance.
Another robot hospitality video came out this week, showcasing Google DeepMind and Stanford Universitys Mobile Aloha a robot that can cook you dinner and then clean up afterward. Researchers claim it only took 50 demonstrations for the robot to understand some new tasks, showing footage of it cooking shrimp and chicken dishes, taking an elevator, opening and closing a cabinet, wiping up some wine and pushing in some chairs. Both the hardware and the machine learning algo are open sourced, and the system costs $20,000 from Trossen Robotics.
Introducing — Hardware!A low-cost, open-source, mobile manipulator.One of the most high-effort projects in my past 5yrs! Not possible without co-lead @zipengfu and @chelseabfinn.At the end, what's better than cooking yourself a meal with the pic.twitter.com/iNBIY1tkcB
— Tony Z. Zhao (@tonyzzhao) January 3, 2024
Full-scale AI plagiarism war
One of the weirder discoveries of the past few months is that Ivy League colleges in the U.S. care more about plagiarism than they do about genocide. This, in a roundabout way, is why billionaire Bill Ackman is now proposing using AI to conduct a plagiarism witch hunt across every university in the world.
Ackman was unsuccessful in his campaign to get Harvard President Claudine Gay fired over failing to condemn hypothetical calls for genocide, but the subsequent campaign to get her fired over plagiarism worked a treat. However, it blew back on his wife Neri Oxman, a former Massachusetts Institute of Technology professor, when Business Insider published claims her 300-page 2010 dissertation had some plagiarised paragraphs.
Read also Features The Road to Bitcoin Adoption is Paved with Whole Numbers Features Why Animism Gives Japanese Characters a NiFTy Head Start on the Blockchain
Ackman now wants to take everyone else down with Neri, starting with a plagiarism review of every academic, admin staff and board member at MIT.“Every faculty member knows that once their work is targeted by AI, they will be outed. No body of written work in academia can survive the power of AI searching for missing quotation marks, failures to paraphrase appropriately, and/or the failure to properly credit the work of others.
Last night, no one at @MIT had a good nights sleep.Yesterday evening, shortly after I posted that we were launching a plagiarism review of all current MIT faculty, President Kornbluth, members of MITs administration, and its board, I am sure that an audible collective gasp
— Bill Ackman (@BillAckman) January 7, 2024
Ackman then threatened to do the same at Harvard, Yale, Princeton, Stanford, Penn and Dartmouth and then surmised that sooner or later, every higher education institution in the world will need to conduct a preemptive AI review of its faculty to get ahead of any possible scandals.
Showing why Ackman is a billionaire and youre not, halfway through his 5,000-word screed, he realized theres money to be made by starting a company offering credible, third-party AI plagiarism reviews and added hed be interested in investing in one.
Enter convicted financial criminal Martin Shkreli. Better known as Pharma Bro for buying the license to Daraprim and then hiking the price by 5,455%, Shkreli now runs a medical LLM service called Dr Gupta. He replied to Ackman, saying: “Yeah I could do this easily,” noting his AI has already been trained on the 36 million papers contained in the PubMed database.
Probably better to do vector search tbh but some rag to go through it case by case. We actually already have all of pubmed downloaded so it's more or less a for loop once you feel good about your main tool.Can calibrate it on Gay's work
— Martin Shkreli (e/acc) (@wagieeacc) January 7, 2024
While online plagiarism detectors like Turnitin already exist, there are doubts about the accuracy and it would still be a mammoth undertaking to enter every article from every academic at even a single institution and cross-check the citations. However, AI agents could potentially systematically and affordably conduct such a review.Even if the global plagiarism witchhunt doesnt happen, it seems increasingly likely that in the next couple of years, any academic who has ever plagiarized something will get found out in the course of a job interview process or whenever they tweet a political position someone else doesnt like.
AI will similarly lower the cost and resource barriers to other fishing expeditions and make it feasible for tax departments to send AI agents to trawl through the blockchain for crypto transactions from 2014 that users failed to report and for offense archaeologists to use AI to comb through every tweet youve made since 2007 looking for inconsistencies or bad takes. Its a brave new world of AI-powered dirt digging.
Two perspectives on AI regulation
Professor Toby Walsh, the chief scientist at the University of New South Waless AI Institute, says heavy-handed approaches to AI regulation are not feasible. He says attempts to limit access to AI hardware like GPUs will not work as LLM compute requirements are falling (see our piece on a local LLM on an iPhone below). He also argues that banning the tech will be about as successful as the United States governments failed efforts to limit access to encryption software in the 1990s.
Read also Features Justin Aversano makes a quantum leap for NFT photography Features North Korean crypto hacking: Separating fact from fiction
Instead he called for vigorous enforcement of existing laws around product liability to hold AI companies to account for the actions of their LLMs. Walsh also called for a focus on competition by applying antitrust regulation more forcibly to lever power away from the Big Tech monopolies, and he called for more investment by governments in AI research.Meanwhile, venture capital firm Andreessen Horowitz has gone hard on the competition is good for AI theme in a letter sent to the United Kingdom’s House of Lords. It says that large AI companies and startups should be allowed to build AI as fast and aggressively as they can and that open-source AI should also be allowed to freely proliferate to compete with both.
A16z’s letter to the House of Lords. (X) All killer, no filler AI news
OpenAI has published a response to The New York Times copyright lawsuit. It claims training on NYT articles is covered by fair use, regurgitation is a rare bug, and despite the NYT case having no merit, OpenAI wants to come to an agreement anyway.
The New Year began with controversy about why ChatGPT is delighted to provide Jewish jokes and Christian jokes but refuses point blank to make Muslim jokes. Someone eventually got a halal-rious pun from ChatGPT, which showed why its best not to ask ChatGPT to make any jokes about anything.
In a development potentially worse than fake kidnappings, AI robocall services have been released that can tie you up in fake spam conversations for hours. Someone needs to develop an AI answering machine to screen these calls.
Introducing Bland Turbo. The world's fastest conversational AI:– Send or receive up to 500,000+ phone calls simultaneously. – Responds at human level speed. In anyone's voice. – Program it to do anything.Call and talk to Bland Turbo now: https://t.co/qm0sUggov3 (1/3) pic.twitter.com/5e6Y3FU9Wh
— Bland.ai (@usebland) January 5, 2024
A blogger with a mixed record claims to have insider info that a big Siri AI upgrade will be announced at Apples 2024 Worldwide Developer Conference. Siri will use the Ajax LLM, resulting in more natural conversations. It will also link to various external services. But who needs Siri AI when you can now download a $1.99 app from the App Store that runs the open-source Mistral 7B 0.2 LLM locally on your iPhone?
Around 170 of the 5,000 submissions to an Australian senate inquiry on legalizing cannabis were found to be AI-generated.
More than half (56%) of 800 CEOs surveyed believe AI will entirely or partially replace their roles. Most also believe that more than half of entry-level knowledge worker jobs will be replaced by AI and that nearly half the skills in the workforce today wont be relevant in 2025.
Subscribe The most engaging reads in blockchain. Delivered once a week.
Email address
SUBSCRIBE