Logs
Recent Search Logs
Query: text classification datasets
Timestamp: 2024-09-18T05:58:14.675207
Results:
[{"title": "One Embedder, Any Task: Instruction-Finetuned Text Embeddings", "url": "https://arxiv.org/abs/2212.09741"}, {"title": "Universal Language Model Fine-tuning for Text Classification", "url": "https://arxiv.org/abs/1801.06146"}, {"title": "Text Embeddings by Weakly-SupervisedContrastive Pre-training", "url": "https://arxiv.org/abs/2212.03533"}]
Query: how to make llm plastic
Timestamp: 2024-09-18T05:57:13.870566
Results:
[{"title": "LLM+P: Empowering Large Language Models with Optimal Planning Proficiency", "url": "https://arxiv.org/abs/2304.11477v1"}, {"title": "Causal Reasoning and Large Language Models:Opening a New Frontier for Causality", "url": "https://arxiv.org/abs/2305.00050"}, {"title": "EnvGen:\nGenerating and Adapting Environments via LLMs for Training Embodied Agents", "url": "https://arxiv.org/abs/2403.12014"}]
Query: LLM as ranker
Timestamp: 2024-09-18T05:56:19.422512
Results:
[{"title": "LLM-Blender: Ensembling Large Language Modelswith Pairwise Ranking and Generative Fusion", "url": "https://arxiv.org/abs/2306.02561"}, {"title": "Lightweight reranking for language model generations", "url": "https://arxiv.org/abs/2307.06857"}, {"title": "Large Language Models as Batteries-Included Zero-Shot ESCO Skills Matchers", "url": "https://arxiv.org/abs/2307.03539"}]
Query: llm knowledge facts
Timestamp: 2024-09-17T14:50:50.891474
Results:
[{"title": "Connecting the Dots: LLMs can Infer and Verbalize Latent Structure from Disparate Training Data", "url": "https://arxiv.org/abs/2406.14546"}, {"title": "SelfCheckGPT: Zero-Resource Black-Box Hallucination Detectionfor Generative Large Language Models", "url": "https://arxiv.org/abs/2303.08896"}, {"title": "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models", "url": "https://arxiv.org/abs/2309.03883"}]
Query: Voice activity projection
Timestamp: 2024-09-17T13:49:10.172829
Results:
[{"title": "Faster Neighborhood Attention:Reducing the\ud835\udcaa\u200b(n2)\ud835\udcaasuperscript\ud835\udc5b2\\mathcal{O}(n^{2})Cost of Self Attention at the Threadblock Level", "url": "https://arxiv.org/abs/2403.04690"}, {"title": "Large Language Models are Few-ShotHealth Learners", "url": "https://arxiv.org/abs/2305.15525"}, {"title": "Distil-Whisper: Robust KnowledgeDistillation via Large-ScalePseudo Labelling", "url": "https://arxiv.org/abs/2311.00430"}]
Query: synthetic data generation for AI text detection
Timestamp: 2024-09-17T13:47:41.665250
Results:
[{"title": "Paraphrasing evades detectors of AI-generated text,but retrieval is an effective defense", "url": "https://arxiv.org/abs/2303.13408"}, {"title": "Best Practices and Lessons Learned on Synthetic Data for Language Models", "url": "https://arxiv.org/abs/2404.07503"}, {"title": "Testing Of Detection Tools For Ai-Generated Text 1", "url": "https://arxiv.org/abs/2306.15666"}]
Query: synthetic data generation for llm text detection
Timestamp: 2024-09-17T13:46:58.852769
Results:
[{"title": "Detection and Measurement of Syntactic Templates in Generated Text", "url": "https://arxiv.org/abs/2407.00211"}, {"title": "Scaling Synthetic Data Creation with 1,000,000,000 Personas", "url": "https://arxiv.org/abs/2406.20094"}, {"title": "Large Language Models as Batteries-Included Zero-Shot ESCO Skills Matchers", "url": "https://arxiv.org/abs/2307.03539"}]
Query: Abstraction and Reasoning Corpus
Timestamp: 2024-09-17T13:44:48.067654
Results:
[{"title": "CodeIt: Self-Improving Language Models with Prioritized Hindsight Replay", "url": "https://arxiv.org/abs/2402.04858"}, {"title": "Neural Abstract Reasoner", "url": "https://arxiv.org/abs/2011.09860"}, {"title": "RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval", "url": "https://arxiv.org/abs/2401.18059"}]
Query: Audio llm
Timestamp: 2024-09-17T13:44:19.827260
Results:
[{"title": "Tart: A plug-and-play Transformer module for task-agnostic reasoning", "url": "https://arxiv.org/abs/2306.07536"}, {"title": "ChatMusician: Understanding and Generating MusicIntrinsically with LLM", "url": "https://arxiv.org/abs/2402.16153"}, {"title": "CoDi-2: In-Context, Interleaved, and Interactive Any-to-Any Generation", "url": "https://arxiv.org/abs/2311.18775"}]
Query: Voice activity projection
Timestamp: 2024-09-17T13:43:54.937858
Results:
[{"title": "Faster Neighborhood Attention:Reducing the\ud835\udcaa\u200b(n2)\ud835\udcaasuperscript\ud835\udc5b2\\mathcal{O}(n^{2})Cost of Self Attention at the Threadblock Level", "url": "https://arxiv.org/abs/2403.04690"}, {"title": "Large Language Models are Few-ShotHealth Learners", "url": "https://arxiv.org/abs/2305.15525"}, {"title": "Distil-Whisper: Robust KnowledgeDistillation via Large-ScalePseudo Labelling", "url": "https://arxiv.org/abs/2311.00430"}]
Query: LLM watermarking
Timestamp: 2024-09-17T13:42:55.173049
Results:
[{"title": "On the Reliability of Watermarks for Large Language Models", "url": "https://arxiv.org/abs/2306.04634"}, {"title": "Can AI-Generated Text be Reliably Detected?", "url": "https://arxiv.org/abs/2303.11156v1"}, {"title": "1 Introduction", "url": "https://arxiv.org/abs/2307.15593"}]
Query: RAG over images
Timestamp: 2024-09-17T13:41:13.600869
Results:
[{"title": "Can Long-Context Language Models Subsume Retrieval, RAG, SQL, and More?", "url": "https://arxiv.org/abs/2406.13121"}, {"title": "The Power of Noise: Redefining Retrieval for RAG Systems", "url": "https://arxiv.org/abs/2401.14887"}, {"title": "Benchmarking Retrieval-Augmented Generation for Medicine", "url": "https://arxiv.org/abs/2402.13178"}]
Query: LLM watermarking
Timestamp: 2024-09-17T13:37:22.619395
Results:
[{"title": "On the Reliability of Watermarks for Large Language Models", "url": "https://arxiv.org/abs/2306.04634"}, {"title": "Can AI-Generated Text be Reliably Detected?", "url": "https://arxiv.org/abs/2303.11156v1"}, {"title": "1 Introduction", "url": "https://arxiv.org/abs/2307.15593"}]
Query: Abstraction and Reasoning Corpus
Timestamp: 2024-09-17T13:36:58.226962
Results:
[{"title": "CodeIt: Self-Improving Language Models with Prioritized Hindsight Replay", "url": "https://arxiv.org/abs/2402.04858"}, {"title": "Neural Abstract Reasoner", "url": "https://arxiv.org/abs/2011.09860"}, {"title": "RAPTOR: Recursive Abstractive Processing for Tree-Organized Retrieval", "url": "https://arxiv.org/abs/2401.18059"}]
Query: RAG over images
Timestamp: 2024-09-17T13:32:24.997201
Results:
[{"title": "Can Long-Context Language Models Subsume Retrieval, RAG, SQL, and More?", "url": "https://arxiv.org/abs/2406.13121"}, {"title": "The Power of Noise: Redefining Retrieval for RAG Systems", "url": "https://arxiv.org/abs/2401.14887"}, {"title": "Benchmarking Retrieval-Augmented Generation for Medicine", "url": "https://arxiv.org/abs/2402.13178"}]
Query: energy based methods
Timestamp: 2024-09-17T11:30:55.836271
Results:
[{"title": "How to Train Your Energy-Based Models", "url": "https://arxiv.org/abs/2101.03288"}, {"title": "Learning Iterative Reasoning through Energy Diffusion", "url": "https://arxiv.org/abs/2406.11179"}, {"title": "Energy Transformer", "url": "https://arxiv.org/abs/2302.07253"}]
Query: LLM
Timestamp: 2024-09-17T11:30:41.777200
Results:
[{"title": "Judging LLM-as-a-Judgewith MT-Bench and Chatbot Arena", "url": "https://arxiv.org/abs/2306.05685"}, {"title": "LLM-Pruner: On the Structural Pruningof Large Language Models", "url": "https://arxiv.org/abs/2305.11627"}, {"title": "LLM+P: Empowering Large Language Models with Optimal Planning Proficiency", "url": "https://arxiv.org/abs/2304.11477v1"}]