MENU
Yu Gu, Sheng Zhang, Naoto Usuyama, Yonas Woldesenbet, Cliff Wong, Praneeth Sanapathi, Mu Wei, Naveen Valluri, Erika Strandberg, Tristan Naumann and Hoifung Poon
Benjamin Grimmer establishes provably faster convergence rates for gradient descent in smooth convex optimization via a computer-assisted analysis technique.
Genomic time-series from experimental evolution studies and ancient DNA datasets offer us a chance to more directly observe the interplay of various evolutionary forces. Alexis Simon and Graham Coop show how the genome-wide variance in allele frequency change between two time points can be decomposed into the contributions of...
Engineers and researchers from Yandex Research, HSE University, University of Washington, Hugging Face, ENS Paris-Saclay, and Yandex School of Data Analysis propose Petals, an open-source decentralized system (showcased this week at the ACL 2023 Demonstrations track) allowing anybody to run large models or even adapt them us...
On the basis of the new adequacy decision, personal data can flow safely from the EU to US companies participating in the Framework, without having to put in place additional data protection safeguards.
Jeremy Howard sheds interesting light on the AI Safety debate
Quantamagazine - Mathematicians can often figure out what happens as quantities grow infinitely large. What about when they are just a little big?
Nelson F. Liu, Kevin Lin, John Hewitt, Ashwin Paranjape, Michele Bevilacqua, Fabio Petroni and Percy Liang find that performance is often highest when relevant information occurs at the beginning or end of the input context, and significantly degrades when models must access relevant information in the middle of long contexts.
Jiayu Ding, Shuming Ma, Li Dong, Xingxing Zhang, Shaohan Huang, Wenhui Wang, Nanning Zheng and Furu Wei introduce LongNet, a Transformer variant that can scale sequence length to more than 1 billion tokens, without sacrificing the performance on shorter sequences.
Quantamagazine - New research reveals how marine microbes use an extra membrane that once had digestive functions to boost their yield from photosynthesis
Nature - R. Z. Moger-Reischer, J. I. Glass, K. S. Wise, L. Sun, D. M. C. Bittencourt, B. K. Lehmkuhl, D. R. Schoolmaster Jr, M. Lynch and J. T. Lennon report on how an engineered minimal cell contends with the forces of evolution compared with the Mycoplasma mycoides non-minimal cell from which it was synthetically derived.
Semianalysis detail the bottlenecks to production and how much downstream capacity is expanding for Nvidia and their competitors
Luciano Del Corro, Allie Del Giorno, Sahaj Agarwal, Bin Yu, Ahmed Awadallah and Subhabrata Mukherjee propose a simple and effective token-level early exit method, SkipDecode, designed to work seamlessly with batch inferencing and KV caching. It overcomes prior constraints by setting up a singular exit point for every token in...
Maanak Gupta, CharanKumar Akiri, Kshitiz Aryal, Eli Parker and Lopamudra Praharaj highlight the limitations, challenges, potential risks, and opportunities of GenAI in the domain of cybersecurity and privacy.
Nature - Cheng Kai Lim, Jing Wui Yeoh, Aurelius Andrew Kunartama, Wen Shan Yew and Chueh Loo Poh detail a method of capturing 2-dimensional light patterns into DNA, by utilizing optogenetic circuits to record light exposure into DNA, encoding spatial locations with barcoding, and retrieving stored images via high-throughput n...
Jared Palmer (Vercel) gives high-level insights into a toolbox to protect a platform against scrapping
Fantastic project from the Princeton University, releasing a whole-brain connectome of the fruit fly, including c.130k annotated neurons and tens of millions of typed synapses!
By Woosuk Kwon, Zhuohan Li, Siyuan Zhuang, Ying Sheng, Lianmin Zheng, Cody Yu, Joey Gonzalez, Hao Zhang and Ion Stoica introduce vLLM, an open-source library for fast LLM inference and serving. vLLM utilizes PagedAttention, a new attention algorithm that effectively manages attention keys and values.
Shen, Zejiang, Ruochen Zhang, Melissa Dell, Benjamin Lee, Jacob Carlson, and Weining Li introduce LayoutParser, an open-source library for streamlining the usage of DL in document image analysis research and applications.
Joint report from Georgetown University’s Center for Security and Emerging Technology (CSET) and The Alan Turing Institute’s Centre for Emerging Technology and Security (CETaS) assesses the current state-of-the-art in autonomous cyber defence and its future potential, identifies barriers to progress and recommends specific ac...
Guillaume Sanchez, Honglu Fan, Alexander Spangher, Elad Levi, Pawan Sasanka Ammanamanchi and Stella Biderman demonstrate that Classifier-Free Guidance can be used broadly as an inference-time technique in pure language modeling.
Positive Security showcase an attack which leverages indirect prompt injection to trick Auto-GPT (LLM-based agent) into executing arbitrary code when it is asked to perform a seemingly harmless task such as text summarization on an attacker controlled website
Li Chen, Penghao Wu, Kashyap Chitta, Bernhard Jaeger, Andreas Geiger and Hongyang Li provide a comprehensive analysis of more than 250 papers, covering the motivation, roadmap, methodology, challenges, and future trends in end-to-end autonomous driving.
Manli Shu, Jiongxiao Wang, Chen Zhu, Jonas Geiping, Chaowei Xiao and Tom Goldstein investigate the exploitability of instruction tuning in large language models. They propose AutoPoison, an automated data poisoning pipeline, to achieve specific attack goals through intentionally changing the model's behavior. The study showca...
Isaac Kauvar, Chris Doyle, Linqi Zhou and Nick Haber present Curious Replay - a form of prioritized experience replay tailored to model-based agents through use of a curiosity-based priority signal.