bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872


1/11
@AnthropicAI
With styles, you can now customize how Claude responds.

Select from the new preset options: Concise, Explanatory, or Formal.



https://video.twimg.com/ext_tw_video/1861472528524337152/pu/vid/avc1/1920x1080/zvAWsXTu5K5R0A6N.mp4

2/11
@AnthropicAI
Want Claude to more closely match how you communicate?

Upload writing samples and Claude can automatically generate custom styles, just for you.



https://video.twimg.com/ext_tw_video/1861472551139934208/pu/vid/avc1/1920x1080/HHCgSB3ATOqAUssN.mp4

3/11
@testingcatalog
Finally someone released something to really improve writing πŸ’ͺ



4/11
@KinggZoom
When will Claude get internet access and better rate limits?



5/11
@thegenioo
Anthropic takes it again OpenAI losing πŸ₯Έ



6/11
@omarsar0
Curious how this works together with Projects? I find myself often using Projects for styled responses.



7/11
@novocrypto
Can there be a β€œprefer not to choose style” option that defaults to Claude as it is? That would be dope 🫑



8/11
@JohnForrester
rather than a menu selection, I think it is better for you to build a default personalization of user styles. Asking users for writing samples and remembering. btw, what is going on with your latency via the API? huge spike in Sonnet latency last 30 days.



9/11
@AGI_Odyssey
Fabulous!



10/11
@koltregaskes
Thank you again guys!



11/11
@alikayadibi11
Good job




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872





1/5
@rohanpaul_ai
Adding rule-based guidance doubles RAG's performance in document retrieval and answer generation.

Basically, RAG gets a proper manual on how to use its knowledge.

It's like giving RAG a GPS instead of letting it wander around blindly.

🎯 Original Problem:

Current Retrieval-Augmented Generation (RAG) frameworks face two major limitations: retrievers can't guarantee fetching the most relevant information, and LLMs lack specific guidance on using retrieved content effectively.

-----

πŸ”§ Solution in this Paper:

β†’ Introduces RuleRAG, which uses symbolic rules to guide both retrieval and generation processes.

β†’ Guide retrievers to fetch logically related documents following rule directions

β†’ Help generators uniformly generate answers attributed by the same set of rules

β†’ Use queries and rules combined as supervised fine-tuning data to improve rule-based instruction following

β†’ RuleRAG-ICL: Uses in-context learning with rule guidance during retrieval and inference

β†’ RuleRAG-FT: Fine-tunes both retrievers and generators using rule-guided fine-tuning

β†’ Created five rule-aware QA benchmarks (three temporal, two static) to evaluate performance

-----

πŸ’‘ Key Insights:

β†’ Rules can explicitly guide both document retrieval and answer generation

β†’ Combining rules with queries improves retrieval quality significantly

β†’ Rule-guided fine-tuning enhances both retrieval and generation performance

β†’ The method scales well with increasing numbers of retrieved documents

-----

πŸ“Š Results:

β†’ RuleRAG-ICL improved retrieval quality by +89.2% in Recall@10 scores

β†’ Generation accuracy increased by +103.1% in exact match scores

β†’ RuleRAG-FT achieved even better performance improvements across all benchmarks

β†’ Method showed strong generalization ability for untrained rules



GdVhJdWaoAghtFJ.png


2/5
@rohanpaul_ai
Paper Title: "RuleRAG: Rule-guided retrieval-augmented generation with language models for question answering"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861501768653512707/pu/vid/avc1/1080x1080/0WhCQR7B3zGenfUy.mp4

3/5
@rohanpaul_ai
Guided by the rule r related to the query, the proposed RuleRAG first retrieves supportive documents that are logically related to the query and then attributes the correct answer,



GdVin-GaoAAkUXI.png


4/5
@rohanpaul_ai




GdVjGAgaoAMRTnf.jpg


5/5
@rohanpaul_ai
πŸ“š [2410.22353] RuleRAG: Rule-guided retrieval-augmented generation with language models for question answering




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872



1/3
@rohanpaul_ai
Efficient bandit-based approach reduces human annotation costs in LLM training

SEA (Sample-Efficient Alignment) uses Thompson sampling to align LLMs with minimal human feedback

🎯 Original Problem:

Aligning LLMs with human preferences requires massive amounts of human feedback data, making it expensive and time-consuming. Current methods need extensive human annotations for effective alignment.

-----

πŸ”§ Solution in this Paper:

β†’ They frame LLM alignment as a contextual dueling bandits problem, where the model learns from pairwise comparisons of responses

This formulation naturally requires two key properties for sample-efficient alignment:

β†’ Online interaction - allowing the agent to act with latest learned policy and immediately improve from experience

β†’ Active exploration - strategically selecting actions that lead to maximal policy improvement

β†’ Introduce SEA (Sample-Efficient Alignment), implementing Thompson sampling with epistemic reward modeling

β†’ SEA maintains uncertainty-aware reward models and uses policy-guided search for efficient exploration

β†’ The system works in both online user feedback and crowdsourcing scenarios

-----

πŸ’‘ Key Insights:

β†’ Online interaction allows immediate policy improvement from latest experiences

β†’ Active exploration strategically selects actions for maximal learning

β†’ Thompson sampling naturally balances exploration vs exploitation

β†’ Mixed preference learning combines different alignment approaches

-----

πŸ“Š Results:

β†’ Tested across three model scales: 1B, 2.8B, and 6.9B parameters

β†’ Evaluated with three preference learning algorithms: DPO, IPO, and SLiC

β†’ Achieved higher win rates compared to baseline approaches

β†’ Significantly improved sample efficiency over recent active exploration methods



GdVdR0WaoAAvrKF.png


2/3
@rohanpaul_ai
Paper Title: "Sample-Efficient Alignment for LLMs"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861496636813516800/pu/vid/avc1/1080x1080/1eUeBGlemWFkbEbV.mp4

3/3
@rohanpaul_ai
[2411.01493] Sample-Efficient Alignment for LLMs




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872

1/1
@rohanpaul_ai
A group appears to have leaked access to Sora, OpenAI’s video generator.

The group created a HuggingFace space that seemingly connects with Sora API, which isn’t yet publicly available.

They built a frontend using their early access tokens, enabling users to create Sora-generated videos.

That HuggingFace space can generate 10-second video clips at 1080p resolution, depending on compute availability.

[Quoted tweet]
OpenAI Sora has leaked


GdU5IfvbgAA9n27.png

GdUhsttW8AEHr-r.png



To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872

1/1
@ccamp___
I am seeing two approaches to social agents: basing them off of real people, and letting them create their own selves.

Re: agents as real people, approaches have been both to train a custom model or give information in prompt context, some examples:

- β€œReid GPT” went viral with Reid Hoffman training a custom GPT off of all his public works
- The @joon_s_pk Generative Agents Simulation paper predicts people's decisions by interviewing them and injecting interviews into prompt context to get the LLM to simulate the interviewee’s choices (would love to see someone train a model on actual vs. predicted results)

For agents as real people, personally I’d love to be able to make a digital twin of myself that takes all of my texts, emails, Zoom calls, everything possible and trains a model on my entire digital footprint. Something kind of like @withdelphi but automatic.

Re: agents as their own people, the approaches all tackle the question of, what makes an agent feel like a real human? Examples:
- Give agents their own experience of being-in-the-world like real humans, with access to news, embodiment, other people/agents, etc. to create their own memories and attitudes
- Improve reasoning, mostly done by foundation models but also things like @NousResearch Forge Reasoning
- Improve their memory systems, like @MemGPT, @mem0ai
- Give them their own goals to pursue, like our real friends and colleagues

The point of all these being, when we talk to a friend, they’re not just trapped away in a box for us to interact with at our leisure, they have their own relationships they learn from, goals, context of the world that makes them interesting.



GdVjTQrWUAEuEwP.png



To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872

1/1
@rohanpaul_ai
πŸš€ Thomson Reuters integrates o1-mini model, into its CoCounsel, achieves 1400% user growth in legal AI

Combines multiple AI models to catch critical legal details other systems miss and teaches AI to think like different types of lawyers

Thomson Reuters integrates OpenAI's o1-mini model into CoCounsel legal assistant, marking first enterprise customization of the model. Implementation showcases strategic use of multiple AI models for specific legal tasks, resulting in 1,400% user growth.

β†’ Thomson Reuters deploys specialized AI models from OpenAI, Google, and Anthropic, each optimized for specific legal workflows. OpenAI handles generative tasks, Google's Gemini processes long-context documents, while Anthropic's Claude manages sensitive tax and compliance cases. The o1-mini model shows superior performance in detecting nuanced legal privileges compared to GPT-4.

β†’ Company expands beyond AI usage through Safe Sign Technologies acquisition for proprietary LLM development. Infrastructure support comes from AWS Sagemaker HyperPod.

This implementation signals a shift from general-purpose AI to precision-engineered specialized models working together. For legal industry, where precision is crucial, this approach could set new standards in AI deployment and error reduction.



GdUXx8haoAMBhiz.jpg



To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872


1/3
@rohanpaul_ai
Speculative decoding just landed in llama.cpp's server πŸ’‘
β†’ Speedups in token generation ranging from 1.25x to 5.73x depending on hardware and model configurations

β†’ The new speculative decoding feature allows using a smaller model to draft tokens which are then verified by the main larger model to predict tokens speculatively.

β†’ Draft model predictions are verified by the main model before being accepted

β†’ Resulting in great speedups in token generation while maintaining the quality of the larger model's outputs

⚠️ Key Considerations

β†’ Performance gains vary based on hardware architecture and memory bandwidth

β†’ Apple Silicon shows less dramatic improvements due to shared memory architecture



GdThtNDagAAA1R6.png


2/3
@rohanpaul_ai
server : add speculative decoding support by ggerganov Β· Pull Request #10455 Β· ggerganov/llama.cpp



3/3
@victor_explore
llamas are getting faster at predicting the future




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872



1/3
@rohanpaul_ai
Finally a model that understands social networks form through both birds-of-a-feather and popularity

NMM ((Non-Euclidean Mixture Model) combines spherical and hyperbolic spaces to better model how social networks actually form

🎯 Original Problem:

Social networks form links through two key factors - homophily (similar nodes connecting) and social influence (popular nodes attracting connections). Current models either focus on just one factor or use simple Euclidean spaces that can't capture the complex network structures.

-----

πŸ”§ Solution in this Paper:

β†’ Introduces NMM (Non-Euclidean Mixture Model) that represents nodes in both spherical space (for homophily) and hyperbolic space (for social influence)

β†’ Uses spherical space to model cyclic connections between similar nodes and hyperbolic space to model hierarchical influence-based connections

β†’ Combines these two representations through a novel space unification loss that aligns the two geometric spaces

β†’ Enhances NMM with graph neural networks (NMM-GNN) using a variational autoencoder framework with non-Euclidean encoders and decoders

-----

πŸ’‘ Key Insights:

β†’ Social network links form through both similarity and influence - need both spherical and hyperbolic spaces to model this effectively

β†’ Homophily creates cycles best captured in spherical space while influence creates hierarchies best captured in hyperbolic space

β†’ Unifying different geometric spaces through projection and alignment is key for coherent node representations

-----

πŸ“Š Results:

β†’ Significantly outperforms state-of-the-art baselines on social network generation tasks

β†’ Successfully tested on large-scale networks like LiveJournal (4.8M nodes) and Friendster (65.6M nodes)

β†’ More parameter efficient compared to existing models like RaRE



GdQ1pPEaoAMsbjM.png


2/3
@rohanpaul_ai
Paper Title: "Non-Euclidean Mixture Model for Social Network Embedding"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861171591931142146/pu/vid/avc1/1080x1080/-WOGpwh9a824cRFM.mp4

3/3
@rohanpaul_ai
[2411.04876v1] Non-Euclidean Mixture Model for Social Network Embedding




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872




1/4
@rohanpaul_ai
New smart weight-compression technique has arrived to reduce your GPU VRAM requirement.

Squeeze more parameters into your GPU by compressing the wasteful parts of floating-point numbers

NeuZip compresses neural networks by exploiting the low entropy nature of floating-point exponents

🎯 Original Problem:

Training and deploying large neural networks is severely constrained by GPU memory limitations. While model sizes have grown 100x since 2017, GPU memory has only increased 2.5x (from 32GB to 80GB), creating a critical bottleneck for scaling neural networks.

-----

πŸ”§ Solution in this Paper:

β†’ Introduces NeuZip - a novel compression scheme that exploits low entropy in neural network parameters' exponent bits

β†’ Compresses exponent bits losslessly using asymmetric numeral system (ANS) while keeping sign and mantissa bits unchanged

β†’ Implements layer-by-layer compression/decompression during training to avoid creating large buffers

β†’ Compatible with activation checkpointing for additional memory savings

β†’ For inference, introduces lossy compression by truncating mantissa bits while controlling relative weight changes

-----

πŸ’‘ Key Insights:

β†’ Neural network parameters tend to concentrate around zero, making exponent bits highly compressible

β†’ Exponent bits carry only ~3 bits of information despite using 8 bits of storage

β†’ Layer-wise compression enables training without ever fully decompressing the entire network

β†’ Inference tasks can tolerate more aggressive lossy compression compared to training

-----

πŸ“Š Results:

β†’ Reduces Llama-3 8B training memory from 31GB to 16GB with no performance loss

β†’ Enables training 13B parameter models on consumer GPUs (20GB memory)

β†’ For inference, achieves >50% memory reduction while maintaining near-lossless performance

β†’ Outperforms QLoRA and other quantization methods in memory-performance trade-off



GdQ4CwCboAACkXl.png


2/4
@rohanpaul_ai
Paper Title: "NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861174229959614464/pu/vid/avc1/1080x1080/TgngZoqe92dCG6tb.mp4

3/4
@rohanpaul_ai




GdQ4p6OaoAEkbd8.png


4/4
@rohanpaul_ai
πŸ“š [2410.20650] NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872




1/5
@rohanpaul_ai
A single model with multiple experts handles error correction for different input types

NeKo, proposed in this paper, uses specialized experts to fix recognition errors across speech, text and vision tasks

Original Problem πŸ€”:

Building a general-purpose post-recognition error corrector that can handle multiple domains (speech, text, vision) while maintaining high performance across all tasks remains challenging. Current solutions require separate models for each domain, leading to parameter inefficiency.

-----

Solution in this Paper πŸ› οΈ:

β†’ NeKo introduces a task-oriented Mixture-of-Experts (MoE) architecture where experts specialize in specific tasks (speech-to-text, language-to-text, vision-to-text)

β†’ During training, input tokens are routed to both their task-specific expert and the top expert selected by a gating network

β†’ During inference, tokens are routed purely based on router probabilities without task knowledge, enabling zero-shot generalization

β†’ The model replaces standard feedforward blocks with MoE layers, allowing efficient parameter sharing across tasks

-----

Key Insights from this Paper πŸ’‘:

β†’ Task-specific expert assignment during training enables better specialization while maintaining cross-task knowledge sharing

β†’ MoE architecture provides better parameter efficiency compared to having separate models for each task

β†’ Zero-shot generalization is possible by relying on learned routing patterns during inference

-----

Results πŸ“Š:

β†’ 5.0% relative Word Error Rate reduction on Open ASR Leaderboard

β†’ 15.5% to 27.6% relative WER reduction compared to GPT-3.5 and Claude-Opus on zero-shot Hyporadise benchmark

β†’ State-of-the-art results in ASR correction while maintaining competitive performance on grammar and OCR correction tasks



GdQ5stpaoAAkoJw.png


2/5
@rohanpaul_ai
Paper Title: "NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861176047334760449/pu/vid/avc1/1080x1080/1V5wmFt-Uzj8Ueyf.mp4

3/5
@rohanpaul_ai
πŸ” NeKo's architecture

The model replaces standard feedforward blocks with MoE layers. During training, each expert is mapped to a specific task, with input tokens being routed to both their task-specific expert and the top expert selected by the gating network.

During inference, tokens are routed purely based on router probabilities without task knowledge.



GdQ6D3QaoAE0cA1.png


4/5
@rohanpaul_ai
[2411.05945] NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts



5/5
@MiaAI_Builder
I appreciate the idea of using specialized experts to correct errors across different input types, a promising approach to building a general-purpose post-recognition error correction model.




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872




1/6
@rohanpaul_ai
New tests reveal the true effective context limits of leading LLMs.

β†’ GPT-4 performs best at small contexts, Gemini 1.5 Pro excels at longer contexts

β†’ Claude 3.5 Sonnet leads in mid-range contexts (2.5k to 32k tokens)

🎯 Original Problem:

Current benchmarks for evaluating LLMs' long context capabilities are inadequate - they either saturate at perfect scores, test on limited context lengths, or lack granular insights into specific model behaviors.

-----

πŸ”¬ Solution in this Paper:

β†’ Introduced a series of increasingly complex retrieval tasks using synthetic UUID key-value pairs to test 17 leading LLMs

β†’ Created novel "needle threading" tasks where models must follow chains of linked information through contexts up to 900k tokens

β†’ Developed tasks like Single Needle (basic retrieval), Multiple Needles (concurrent retrieval), and Threading (following information chains)

β†’ Introduced Multi-Threading to test if models can track multiple information threads simultaneously

-----

πŸ’‘ Key Insights:

β†’ Most models' effective context limit is shorter than their advertised context length

β†’ Models perform better with forward-moving threads compared to backward threads

β†’ Many models are remarkably "thread-safe" - can follow multiple threads without performance loss

β†’ Different tokenizers count tokens very differently - direct comparisons can be misleading

β†’ Performance generally decreases towards the middle of the context window

-----

πŸ“Š Results:

β†’ GPT-4 performs best at small contexts, Gemini 1.5 Pro excels at longer contexts

β†’ Claude 3.5 Sonnet leads in mid-range contexts (2.5k to 32k tokens)

β†’ Closed-source models consistently outperform open-source alternatives

β†’ Most models show significant accuracy drop beyond their effective context limit



GdRLiEraoAMAoM4.png


2/6
@rohanpaul_ai
Paper Title: "Needle Threading: Can LLMs Follow Threads through Near-Million-Scale Haystacks?"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861195970811445248/pu/vid/avc1/1080x1080/tHjcPg5xbjO0bzhA.mp4

3/6
@rohanpaul_ai
πŸ”¬ Evaluation methods used

β†’ Used 17 leading LLMs including GPT-4, Gemini 1.5, Claude 3, and open-source models

β†’ Tested on contexts ranging from 1k to 630k tokens

β†’ Evaluated using exact matching with expected answers

β†’ Introduced a task-specific effective context limit metric to measure real performance capabilities



GdRMPvHa0AAqSFb.png


4/6
@rohanpaul_ai
[2411.05000] Needle Threading: Can LLMs Follow Threads through Near-Million-Scale Haystacks?



5/6
@tOSUFever
everything every pixel is all in long context.
multithreaded (better) = yes.
gpt4 best at short context = (this is probably wrong) i think gpt4 long context was your grandaddy worldmodel distilling data synthetically initially at great cost.

also we need some new words.

long context serves two purposes the first is prompting with a ton of data like a book or repo. (this is still one turn one prompt), human expects ai output

the second is different; the behavior of a long conversation thread. this isn’t one turn this is thousands of unique turns. this is when you SEE behavior. human INPUT to context, ai (output) is INPUT to context.

and human INPUT (as far as the context window is concerned for β€˜the next output’) is weighted differently.

for example llm providers may not pass through your exact prompt to β€œprevent hallucinations” or β€œjailbreaking”. maybe characters like _* or (…) whatever. however the model will happily insert them for you into context…

and if it *likes* you (literally) the next outputs only get better πŸ’₯



6/6
@medoraai
Anthropic's new enterprise version has a 500K context window. Interested to see how they do. Google is the king here however.




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872



1/3
@rohanpaul_ai
This paper makes complex Multi-objective reinforcement learning (MORL) policies understandable by clustering them based on both behavior and objectives

When AI gives you too many options, this clustering trick saves the day

🎯 Original Problem:

Multi-objective reinforcement learning (MORL) generates multiple policies with different trade-offs, but these solution sets are too large and complex for humans to analyze effectively. Decision makers struggle to understand relationships between policy behaviors and their objective outcomes.

-----

πŸ› οΈ Solution in this Paper:

β†’ Introduces a novel clustering approach that considers both objective space (expected returns) and behavior space (policy actions)

β†’ Uses Highlights algorithm to capture 5 key states that represent each policy's behavior

β†’ Applies PAN (Pareto-Set Analysis) clustering to find well-defined clusters in both spaces simultaneously

β†’ Employs bi-objective evolutionary algorithm to optimize clustering quality across both spaces

-----

πŸ’‘ Key Insights:

β†’ First research to tackle MORL solution set explainability

β†’ Different policies with similar trade-offs can exhibit vastly different behaviors

β†’ Combining objective and behavior analysis reveals deeper policy insights

β†’ Makes MORL more practical for real-world applications

-----

πŸ“Š Results:

β†’ Outperformed traditional k-medoids clustering in MO-Highway and MO-Lunar-lander environments

β†’ Showed comparable performance in MO-Reacher and MO-Minecart scenarios

β†’ Successfully demonstrated practical application through highway environment case study



GdRNNQaaoAAiEE1.jpg


2/3
@rohanpaul_ai
Paper Title: "Navigating Trade-offs: Policy Summarization for Multi-Objective Reinforcement Learning"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861197491238211584/pu/vid/avc1/1080x1080/56yXAj4Toyxny-Ic.mp4

3/3
@rohanpaul_ai
[2411.04784v1] Navigating Trade-offs: Policy Summarization for Multi-Objective Reinforcement Learning




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872




1/6
@rohanpaul_ai
Nice survey paper presents a unified taxonomy bridging personalized text generation and downstream applications

🎯 Current research on LLM personalization is fragmented into two disconnected areas: direct personalized text generation and downstream task personalization.

This split creates a knowledge gap, limiting the development of comprehensive personalization solutions.

This Paper:

β†’ Establishes three personalization granularity levels: user-level (individual), persona-level (groups), and global preference alignment

β†’ Proposes systematic frameworks for personalization techniques including RAG, prompt engineering, fine-tuning, embedding learning, and RLHF

β†’ Creates evaluation taxonomies distinguishing between direct (text quality) and indirect (task performance) assessment methods

-----

πŸ’‘ Key Insights:

β†’ Personalization can be achieved at different granularities, with trade-offs between precision and data requirements

β†’ User-level personalization offers finest control but needs substantial user data

β†’ Persona-level grouping helps handle cold-start problems with new users

β†’ Privacy concerns and bias management are critical challenges

β†’ Multi-modal personalization remains an open challenge



GdQVU-IakAAdNe8.png


2/6
@rohanpaul_ai
Paper Title: "Personalization of Large Language Models: A Survey"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861136074451623938/pu/vid/avc1/1080x1080/ZcMPQJ7Cxpxrf4Pj.mp4

3/6
@rohanpaul_ai
[2411.00027] Personalization of Large Language Models: A Survey



4/6
@rohanpaul_ai




GdQWruxaoAAOv01.png


5/6
@Zhehao_Zhang123
Thanks Rohan for sharing our work!



6/6
@Adina_Coder
Amazing share




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872





1/6
@rohanpaul_ai
Open-source alternative to GPT-4V for building reliable GUI automation agents

OS-ATLAS, a foundational GUI action model enables open-source GUI agents to match commercial VLM performance through cross-platform data synthesis.

Release the largest open-source cross-platform GUI grounding corpus to date, which contains over 13 million GUI elements.

πŸ€– Original Problem:

Existing GUI agents heavily depend on commercial Vision-Language Models (VLMs) like GPT-4V. Open-source VLMs perform poorly in GUI grounding and Out-Of-Distribution scenarios, making them less preferred for real-world applications.

-----

πŸ› οΈ Solution in this Paper:

β†’ Created OS-ATLAS, a foundational GUI action model with three operating modes: Grounding, Action, and Agent

β†’ Built first multi-platform GUI data synthesis toolkit covering Windows, Linux, MacOS, Android, web

β†’ Created largest open-source cross-platform GUI corpus (13M+ elements from 2.3M screenshots)

β†’ Implemented unified action space during training to resolve naming conflicts across platforms

β†’ Standardized Basic Actions (click, type, scroll) and Custom Actions for extensibility

-----

πŸ’‘ Key Insights:

β†’ Pre-training on comprehensive cross-platform GUI data significantly improves grounding accuracy

β†’ Unified action space prevents performance degradation from naming conflicts

β†’ Instruction grounding data, while valuable, isn't critical - referring expression data is sufficient

β†’ Web-only training doesn't generalize well to other platforms

-----

πŸ“Š Results:

β†’ Achieves 82.47% average grounding accuracy without planner

β†’ Reaches 85.14% accuracy with GPT-4 as planner

β†’ Outperforms previous SOTA across mobile, desktop and web platforms

β†’ Shows 14.63% success rate on OSWorld benchmark (compared to 9.21% baseline)



GdQXG9qbkAAUDqR.png


2/6
@rohanpaul_ai
Paper Title: "OS-ATLAS: A Foundation Action Model for Generalist GUI Agents"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861138549690769411/pu/vid/avc1/720x720/oRPlbm8e6lhO-ppC.mp4

3/6
@rohanpaul_ai
The first multi-platform GUI grounding data synthesis toolkit that works across Windows, Linux, MacOS, Android and web platforms



GdQXuyKaoAExE5T.jpg


4/6
@rohanpaul_ai
πŸ“š [2410.23218] OS-ATLAS: A Foundation Action Model for Generalist GUI Agents



5/6
@rohanpaul_ai




GdQX-6IaoAEz-YD.jpg


6/6
@tOSUFever
πŸ”₯




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 

bnew

Veteran
Joined
Nov 1, 2015
Messages
56,193
Reputation
8,249
Daps
157,872





1/7
@rohanpaul_ai
O1 doesn't cheat on math tests - it actually knows how to solve them

A/B testing reveals o1's true mathematical reasoning capabilities beyond memorization

🎯 Original Problem:

OpenAI's Orion-1 (o1) model claims superior reasoning capabilities, but skeptics suggest its performance might stem from memorizing solutions rather than true reasoning abilities.

-----

πŸ”§ Solution in this Paper:

β†’ Used A/B testing comparing o1's performance on two datasets: IMO problems (easily accessible) and CNT problems (less accessible but similar difficulty)

β†’ Implemented a 7-point grading system: 1 point for correct numerical answer, 2 points for intuitive approach, 4 points for detailed reasoning

β†’ Categorized problems into "search" type (finding specific solutions) and "solve" type (equations/optimization)

-----

πŸ’‘ Key Insights:

β†’ O1 shows strong intuitive reasoning and pattern discovery capabilities

β†’ Performs exceptionally well on "search" type problems (~70% accuracy)

β†’ Struggles with rigorous proof steps and "solve" type problems (~21% accuracy)

β†’ Often uses trial-and-error approach instead of formal proofs

-----

πŸ“Š Results:

β†’ No significant performance difference between IMO (51.4%) and CNT (48%) datasets

β†’ T-statistics close to 0, suggesting o1 relies on reasoning rather than memorization

β†’ Outperforms GPT-4o's benchmark of 39.97% on both datasets



GdQyuWIbgAATGrz.png


2/7
@rohanpaul_ai
Paper Title: "OpenAI-o1 AB Testing: Does the o1 model really do good reasoning in math problem solving?"

Generated below podcast on this paper with Google's Illuminate.



https://video.twimg.com/ext_tw_video/1861168366934990851/pu/vid/avc1/1080x1080/D1uoa5hsOde-Fyr9.mp4

3/7
@dikksonPau
It's not Orion-1...



4/7
@rohanpaul_ai
yes the paper refers to o1-preview and o1-mini variants



5/7
@rohanpaul_ai
This paper is actually referring to 01-preview model (not the final 01)



6/7
@rohanpaul_ai
[2411.06198] OpenAI-o1 AB Testing: Does the o1 model really do good reasoning in math problem solving?



7/7
@TShirtnJeans2
Wait, hold on. There are folks who have access to OpenAI's GPT-5/Orion-1 model?

I thought that was scheduled to come out until next year?




To post tweets in this format, more info here: https://www.thecoli.com/threads/tips-and-tricks-for-posting-the-coli-megathread.984734/post-52211196
 
Top