Showing 1–19 of 19 results
/ Date/ Name
Jan 21, 2022AutoDistill: an End-to-End Framework to Explore and Distill Hardware-Efficient Language ModelsJul 24, 2019Benchmarking TPU, GPU, and CPU Platforms for Deep LearningAug 13, 2019Exploiting Parallelism Opportunities with Deep Learning FrameworksAug 29, 2024LLMs generate structurally realistic social networks but overestimate political homophilyJul 31, 2024Gemma 2: Improving Open Language Models at a Practical SizeJul 7, 2025Gemini 2.5: Pushing the Frontier with Advanced Reasoning, Multimodality, Long Context, and Next Generation Agentic CapabilitiesJul 17, 2025Apple Intelligence Foundation Language Models: Tech Report 2025Mar 8, 2024Gemini 1.5: Unlocking multimodal understanding across millions of tokens of contextDec 19, 2023Gemini: A Family of Highly Capable Multimodal ModelsAug 29, 2025Reshaping the Forward-Forward Algorithm with a Similarity-Based ObjectiveFeb 2, 2023Mixed Precision Post Training Quantization of Neural Networks with Sensitivity Guided SearchApr 15, 2026Edge-Side Residual Timing and Frequency Control for Software-Defined Ground Stations in 5G NTN UplinksFeb 23, 2026PuppetChat: Fostering Intimate Communication through Bidirectional Actions and MicronarrativesDec 13, 2021GLaM: Efficient Scaling of Language Models with Mixture-of-ExpertsOct 5, 2023Hadamard Domain Training with Integers for Class Incremental Quantized LearningFeb 10, 2025Machine Learning Fleet Efficiency: Analyzing and Optimizing Large-Scale Google TPU Systems with ML Productivity GoodputNov 7, 2020Exploring the limits of Concurrency in ML Training on Google TPUsJun 8, 2023Augmenting Hessians with Inter-Layer Dependencies for Mixed-Precision Post-Training QuantizationJan 10, 2026Burn-After-Use for Preventing Data Leakage through a Secure Multi-Tenant Architecture in Enterprise LLM