Showing 1–20 of 55 results
/ Date/ Name
Oct 24, 2022NASA: Neural Architecture Search and Acceleration for Hardware Inspired Hybrid NetworksOct 14, 2022An Efficient FPGA Accelerator for Point CloudFeb 3, 2023PDPU: An Open-Source Posit Dot-Product Unit for Deep Learning ApplicationsDec 12, 2018An Active-Passive Measurement Study of TCP Performance over LTE on High-speed RailsSep 21, 2024SPEED: A Scalable RISC-V Vector Processor Enabling Efficient Multi-Precision DNN InferenceSep 26, 2024Efficient Arbitrary Precision Acceleration for Large Language Models on GPU Tensor CoresMay 25, 2025Enable Lightweight and Precision-Scalable Posit/IEEE-754 Arithmetic in RISC-V Cores for Transprecision ComputingJul 16, 2024Co-Designing Binarized Transformer and Hardware Accelerator for Efficient End-to-End Edge DeploymentMay 25, 2025FastMamba: A High-Speed and Efficient Mamba Accelerator on FPGA with Accurate QuantizationNov 3, 2025Memory-Efficient Training with In-Place FFT ImplementationNov 21, 2024TaQ-DiT: Time-aware Quantization for Diffusion TransformersSep 18, 2020Hardware Accelerator for Multi-Head Attention and Position-Wise Feed-Forward in the TransformerSep 6, 2019Training Deep Neural Networks Using Posit Number SystemJun 24, 2021Transform-Based Feature Map Compression for CNN InferenceMay 8, 2019A Hardware-Oriented and Memory-Efficient Method for CTC DecodingOct 18, 2022Accelerate Three-Dimensional Generative Adversarial Networks Using Fast AlgorithmAug 16, 2023S2R: Exploring a Double-Win Transformer-Based Framework for Ideal and Blind Super-ResolutionApr 11, 2012Reduced-Complexity Column-Layered Decoding and Implementation for LDPC CodesSep 19, 2024A High-Throughput Hardware Accelerator for Lempel-Ziv 4 Compression AlgorithmMay 6, 2024Trio-ViT: Post-Training Quantization and Acceleration for Softmax-Free Efficient Vision Transformer