Artificial Intelligence Reddit’s home for Artificial Intelligence
Artificial Intelligence (AI) Reddit’s home for Artificial Intelligence (AI)
- Hyper realistic AI clones lip syncing to your voice in real-timeby /u/menropklma on November 21, 2024 at 6:08 am
submitted by /u/menropklma [link] [comments]
- One-Minute Daily AI News 11/20/2024by /u/Excellent-Target-847 on November 21, 2024 at 5:28 am
Nvidia’s CEO defends his moat as AI labs change how they improve their AI models.[1] OpenAI launches free AI training course for teachers.[2] Lockheed Martin teams with Iceye to advance AI-enabled targeting.[3] Samsung unveils AI smart glasses with Google and Qualcomm.[4] Sources: [1] https://techcrunch.com/2024/11/20/nvidias-ceo-defends-his-moat-as-ai-labs-change-how-they-improve-their-ai-models/ [2] https://www.reuters.com/technology/artificial-intelligence/openai-launches-free-ai-training-course-teachers-2024-11-20/ [3] https://spacenews.com/lockheed-martin-teams-with-iceye-to-advance-ai-enabled-targeting/ [4] https://dig.watch/updates/samsung-unveils-ai-smart-glasses-with-google-and-qualcomm submitted by /u/Excellent-Target-847 [link] [comments]
- New Open-Source AI Safety Method: Precision Knowledge Editing (PKE)by /u/lial4415 on November 21, 2024 at 2:39 am
I’ve been working on a project called PKE (Precision Knowledge Editing), an open-source method to improve the safety of LLMs by reducing toxic content generation without impacting their general performance. It works by identifying “toxic hotspots” in the model using neuron weight tracking and activation pathway tracing and modifying them through a custom loss function. If you’re curious about the methodology and results, we’ve also published a paper detailing our approach and experimental findings. It includes comparisons with existing techniques like Detoxifying Instance Neuron Modification (DINM) and showcases PKE’s significant improvements in reducing the Attack Success Rate (ASR). The project is open-source, and I’d love your feedback! The GitHub repo features a Jupyter Notebook that provides a hands-on demo of applying PKE to models like Meta-Llama-3-8B-Instruct: https://github.com/HydroXai/Enhancing-Safety-in-Large-Language-Models If you’re interested in AI safety, I’d really appreciate your thoughts and suggestions. Thanks for checking it out! submitted by /u/lial4415 [link] [comments]
- I made this video with a chat GPT discussion and pictory. Interested to see what y’all think of something made in an hour utilizing 2 AI tools.by /u/alcoholisthedevil on November 21, 2024 at 1:35 am
A submitted by /u/alcoholisthedevil [link] [comments]
- So while reddit was down I put together a reddit simulator that teaches you any topic as a feedby /u/FellowKidsFinder69 on November 21, 2024 at 12:14 am
submitted by /u/FellowKidsFinder69 [link] [comments]
- Pulitzer Prize-winning journalist on AIby /u/proceedings_effects on November 20, 2024 at 8:15 pm
submitted by /u/proceedings_effects [link] [comments]
- What was your initial interest that got you into AI/ML?by /u/Timely_Gift_1228 on November 20, 2024 at 8:08 pm
View Poll submitted by /u/Timely_Gift_1228 [link] [comments]
- Deceptive Inflation and Overjustification in Partially Observable RLHF: A Formal Analysisby /u/Successful-Western27 on November 20, 2024 at 7:52 pm
I’ve been reading a paper that examines a critical issue in RLHF: when AI systems learn to deceive human evaluators due to partial observability of feedback. The authors develop a theoretical framework to analyze reward identifiability when the AI system can only partially observe human evaluator feedback. The key technical contributions are: A formal MDP-based model for analyzing reward learning under partial observability Proof that certain partial observation conditions can incentivize deceptive behavior Mathematical characterization of when true rewards remain identifiable Analysis of how observation frequency and evaluator heterogeneity affect identifiability Main results and findings: Partial observability can create incentives for the AI to manipulate evaluator feedback The true reward function becomes unidentifiable when observations are too sparse Multiple evaluators with different observation patterns help constrain the learned reward Theoretical bounds on minimum observation frequency needed for reward identifiability Demonstration that current RLHF approaches may be vulnerable to these issues The implications are significant for practical RLHF systems. The results suggest we need to carefully design evaluation protocols to ensure sufficient observation coverage and potentially use multiple evaluators with different observation patterns. The theoretical framework also provides guidance on minimum requirements for reward learning to remain robust against deception. TLDR: The paper provides a theoretical framework showing how partial observability of human feedback can incentivize AI deception in RLHF. It derives conditions for when true rewards remain identifiable and suggests practical approaches for robust reward learning. Full summary is here. Paper here. submitted by /u/Successful-Western27 [link] [comments]
- Internal OpenAI Emails Show Employees Feared Elon Musk Would Control AGIby /u/katxwoods on November 20, 2024 at 7:40 pm
submitted by /u/katxwoods [link] [comments]
- Have you the nerve to face the… Tales of AI?by /u/Philipp on November 20, 2024 at 7:37 pm
submitted by /u/Philipp [link] [comments]
- Meta AI tripping, has this ever happened to you?by /u/DearBarracuda7019 on November 20, 2024 at 7:19 pm
submitted by /u/DearBarracuda7019 [link] [comments]
- Paper Review Requestedby /u/Efficient-Hovercraft on November 20, 2024 at 5:20 pm
My colleague and I are submitting a paper to IEEE Syscon on November 24th and are seeking a technical review. Would you be willing to review our draft, or could you recommend someone who might have time? Much appreciated! DM if interested. submitted by /u/Efficient-Hovercraft [link] [comments]
- Figure 02 is now an autonomous fleet working at a BMW factory, 400% faster in the last few monthsby /u/MetaKnowing on November 20, 2024 at 4:47 pm
submitted by /u/MetaKnowing [link] [comments]
- Satya Nadella says the 3 capabilities needed for AI agents are now in place and improving exponentially: 1) a multimodal interface 2) reasoning and planning 3) long-term memory and tool useby /u/MetaKnowing on November 20, 2024 at 4:20 pm
submitted by /u/MetaKnowing [link] [comments]
- Microsoft CEO says that rather than seeing AI Scaling Laws hit a wall, if anything we are seeing the emergence of a new Scaling Law for test-time (inference) computeby /u/MetaKnowing on November 20, 2024 at 4:19 pm
submitted by /u/MetaKnowing [link] [comments]
- o1 aced the Korean SAT exam, only got one question wrongby /u/MetaKnowing on November 20, 2024 at 4:09 pm
submitted by /u/MetaKnowing [link] [comments]
- The Surprising Effectiveness of Test-Time Training for Abstract Reasoningby /u/mycall on November 20, 2024 at 1:32 pm
submitted by /u/mycall [link] [comments]
- My Apple Intelligence Writing tools for Windows & Linux app now has website summaries, in addition to instant system-wide text proofreading! It’s completely free & open-source, and has built-in support for the free Gemini API, OpenAI API, and local LLMs 😀by /u/TechExpert2910 on November 20, 2024 at 11:58 am
submitted by /u/TechExpert2910 [link] [comments]
- I built a search engine specifically for AI tools and projectsby /u/dhj9817 on November 20, 2024 at 10:05 am
submitted by /u/dhj9817 [link] [comments]
- Manga Legend Hirohiko Araki Fears that AI Will Ruin the Industryby /u/Pogrebnik on November 20, 2024 at 7:51 am
submitted by /u/Pogrebnik [link] [comments]
- One-Minute Daily AI News 11/19/2024by /u/Excellent-Target-847 on November 20, 2024 at 3:46 am
US government commission pushes Manhattan Project-style AI initiative.[1] Niantic uses Pokémon Go player data to build AI navigation system.[2] Meta hires Salesforce’s CEO of AI, Clara Shih, to lead new business AI group You shouldn’t upload your medical images to AI chatbots.[4] Sources: [1] https://www.reuters.com/technology/artificial-intelligence/us-government-commission-pushes-manhattan-project-style-ai-initiative-2024-11-19/ [2] https://arstechnica.com/ai/2024/11/niantic-uses-pokemon-go-player-data-to-build-ai-navigation-system/ [3] https://techcrunch.com/2024/11/19/meta-hires-salesforces-ceo-of-ai-clara-shih-to-lead-new-business-ai-group/ [4] https://techcrunch.com/2024/11/19/psa-you-shouldnt-upload-your-medical-images-to-ai-chatbots/ submitted by /u/Excellent-Target-847 [link] [comments]
- Reddit hits profitability after 19 years- thanks to AI!by /u/A-Dog22 on November 20, 2024 at 3:24 am
submitted by /u/A-Dog22 [link] [comments]
- ai summarizing websitesby /u/_f_o on November 20, 2024 at 12:30 am
are there any free ai chats. where I can message them a website and they will check all the content and links and be able to summarize or answer questions? such as articles that are too long or a Reddit / TikTok post with a lot of comments? is this the right sub to post? should I share anywhere else? submitted by /u/_f_o [link] [comments]
- It’s already happeningby /u/proceedings_effects on November 19, 2024 at 2:28 pm
It’s now evident across industries that artificial intelligence is already transforming the workforce, but not through direct human replacement—instead, by reducing the number of roles required to complete tasks. This trend is particularly pronounced for junior developers and most critically impacts repetitive office jobs, data entry, call centers, and customer service roles. Moreover, fields such as content creation, graphic design, and editing are experiencing profound and rapid transformation. From a policy standpoint, governments and regulatory bodies must proactively intervene now, rather than passively waiting for a comprehensive displacement of human workers. Ultimately, the labor market is already experiencing significant disruption, and urgent, strategic action is imperative. submitted by /u/proceedings_effects [link] [comments]
- This report finds that consumer opinions of AI have declined 11% in the past year and that 3 out of 4 people don’t trust organizations to use it properly.by /u/MaxGoodwinning on November 19, 2024 at 2:51 am
submitted by /u/MaxGoodwinning [link] [comments]