ao ren
professor
email: ren.ao@cqu.edu.cn
ao ren obtained his ph.d. degree from the department of electrical and computer engineering, northeastern university, ma, usa. he served as an assistant professor in the department of electrical and computer engineering, clemson university, sc, usa. he is a professor at the college of computer science, chongqing university, china. his research centers around energy-efficient deep learning systems for edge devices, which include deep learning compression algorithms, the deep learning compiler, and the deep learning accelerator architecture for both fpgas and asics. his other research interests include multimodal learning, few-shot learning, and approximate computing. his work has appeared in some of the top venues in computer architecture and artificial intelligence, such as asplos, isca, aaai, isscc, and iccad. he has served as the tpc members of asap, icmla, and iccad.
selected publication (* equal contribution)
conference papers
1. [ijcnn'21] li li, moming duan, duo liu, yu zhang, ao ren, xianzhang chen, yujuan tan and chengliang wang, “fedsae: a novel self-adaptive federated learning framework in heterogeneous systems,” in proc. of the international joint conference on neural networks (ijcnn), 2021. (ccf c)
2. [ijcnn'21] yu zhang, moming duan, duo liu, li li, ao ren, xianzhang chen, yujuan tan and chengliang wang, “fedsae: a novel self-adaptive federated learning framework in heterogeneous systems,” in proc. of the international joint conference on neural networks (ijcnn), 2021. (ccf c)
3. [aaai'20] a. ren, t. zhang, y. wang, s. lin, p. dong, y.k. chen, y. xie, and y. wang, “darb: a density-adaptive regular-block pruning for deep neural networks,” in proc. of the 34th aaai conference on artificial intelligence (aaai), 2020. (ccf a,acceptance rate:20.8%)
4. [asplos'19] a. ren, t. zhang, s. ye, w. xu, x. qian, x. lin, and y. wang, “admm-nn: an algorithm-hardware co-design framework of dnns using alternating direction methods of multipliers,” in proc. of the 24th international conference on architectural support for programming languages and operating systems (asplos), 2019. (ccf a,acceptance rate:21.1%)
5. [isca'19] r. cai, a. ren, o. chen, n. liu, c. ding, x. qian, j. han, w. luo, n. yoshikawa, and y. wang, “a stochastic-computing based deep learning framework using adiabatic quantum-flux-parametron superconducting technology,” in proc. of the 46th international symposium on computer architecture (isca), 2019. (ccf a,acceptance rate:16.9%)
6. [isscc'19] j. yue, r. liu, w. sun, z. yuan, z. wang, y. tu, y. chen, a. ren, y. wang, m. chang, x. li, h. yang, and y. liu, “7.5 a 65nm 0.39-to-140.3 tops/w 1-to-12b unified neural network processor using block-circulant-enabled transpose-domain acceleration with 8.1× higher tops/mm2 and 6t hbst-tram-based 2d data-reuse architecture,” in proc. of ieee international solid-state circuits conference (isscc), 2019. (ccf a)
7. [iccd'19] r. cai, o. chen, a. ren, n. liu, n. yoshikawa, and y. wang, “a buffer and splitter insertion framework for adiabatic quantum-flux-parametron superconducting circuits,” in proc. of the 37th ieee international conference on computer design (iccd), 2019. (ccf b)
8. [glvlsi'19] r. cai, o. chen, a. ren, n. liu, c. ding, n. yoshikawa, and y. wang, “a majority logic synthesis framework for adiabatic quantum-flux-parametron superconducting circuits,” in proc. of the 29th acm great lakes symposium on vlsi (glvlsi), 2019. (ccf c)
9. [isvlsi'19] r. cai, x. ma, o. chen, a. ren, n. liu, n. yoshikawa, and y. wang, “ide development, logic synthesis and buffer/splitter insertion framework for adiabatic quantum-flux-parametron superconducting circuits,” in proc. of ieee computer society annual symposium on vlsi (isvlsi), 2019. (ccf c)
10. [asplos'18,] r. cai, a. ren*, n. liu, c. ding, l. wang, x., m. pedram, and y. wang, “vibnn: hardware acceleration of bayesian neural networks,” in proc. of the 23th international conference on architectural support for programming languages and operating systems (asplos), 2018. (ccf a,acceptance rate:17.6%)
11. [glvlsi'18] c. ding, a. ren*, g. yuan, x. ma, j. li, n. liu, b. yuan, and y. wang, “structured weight matrices-based hardware accelerators in deep neural networks: fpgas and asics,” in proc. of the 28th acm great lakes symposium on vlsi (glvlsi), 2018. (ccf c)
12. [isvlsi'18] z. li, j. li, a. ren, c. ding, j. draper, q. qiu, b. yuan, and y. wang, “towards budget-driven hardware optimization for deep convolutional neural networks using stochastic computing,” in proc. of ieee computer society annual symposium on vlsi (isvlsi), 2018. (ccf c)
13. [isqed'18] x. ma, y. zhang, g. yuan, a. ren, z. li, j. han, j. hu, and y. wang, “an area and energy efficient design of domain-wall memory-based deep convolutional neural networks using stochastic computing,” in proc. of 19th international symposium on quality electronic design (isqed), 2018. best paper nomination
14. [asplos'17] a. ren, j. li, z. li, c. ding, x. qian, q. qiu, b. yuan, and y. wang, “sc-dcnn: highly-scalable deep convolutional neural network using stochastic computing,” in proc. of the 22th international conference on architectural support for programming languages and operating systems (asplos), 2017. (ccf a,acceptance rate:17.4%)
15. [iccad'17] h. li, t. wei, a. ren, q. zhu, and y. wang, “deep reinforcement learning: frame-work, applications, and embedded implementations,” in proc. of the 36th international conference on computer-aided design (iccad), 2017. (ccf b)
16. [iccd'17] r. cai, a. ren, l. wang, m. pedram, and y. wang, “hardware acceleration of bayesian neural networks using ram based linear feedback gaussian random number generators,” in proc. of the 35th ieee international conference on computer design (iccd), 2017. (ccf b)
17. [mwscas'17] g. yuan, c. ding, r. cai, x. ma, z. zhao, a. ren, b. yuan, and y. wang, “memristor crossbar-based ultra-efficient next-generation baseband processors,” in proc. of the 60th ieee international midwest symposium on circuits and systems (mwscas), 2017.
18. [ijcnn'17] j. li, z. yuan, z. li, c. ding, a. ren, q. qiu, j. draper, and y. wang, “hardware-driven nonlinear activation for stochastic computing based deep convolutional neural networks,” in proc. of the international joint conference on neural networks (ijcnn), 2017. (ccf c)
19. [glvlsi'17] z. yuan, j. li, z. li, c. ding, a. ren, b. yuan, q. qiu, j. draper, and y. wang, “softmax regression design for stochastic computing based deep convolutional neural networks,” in proc. of the 27th acm great lakes symposium on vlsi (glvlsi), 2017. (ccf c)
20. [date'17] z. li, a. ren, j. li, q. qiu, b. yuan, j. draper, and y. wang, “structural design optimization for deep convolutional neural networks using stochastic computing,” in proc. of the conference on design, automation and test in europe (date), 2017. (ccf b)
21. [icassp'17] s. liu, a. ren, y. wang, p. k. varshney, “ultra-fast robust compressive sensing based on memristor crossbars,” in proc. of 42nd ieee international conference on acoustics, speech and signal processing (icassp), 2017. best paper award, best student presentation award (top 3 in more than 2,000 submissions) (ccf b)
22. [asp-dac'17] a. ren, s. liu, r. cai, w. wen, p. k. varshney, and y. wang, “algorithm-hardware co-optimization of the memristor-based framework for solving socp and homogeneous qcqp problems,” in proc. of 22nd asia and south pacific design automation conference (asp-dac), 2017.(ccf c)
23. [asp-dac'17] j. li, a. ren, z. li, c. ding, b. yuan, q. qiu, and y. wang, “towards acceleration of deep convolutional neural networks using stochastic computing,” in proc. of 22nd asia and south pacific design automation conference (asp-dac), 2017. (ccf c)
24. [icrc'16] a. ren, z. li, b. yuan, q. qiu, and y. wang, “designing reconfigurable large-scale deep learning systems using stochastic computing,” in proc. of ieee international conference on rebooting computing (icrc), 2016.
25. [iccd'16] zhe li, a. ren*, j. li, q. qiu, y. wang, and b. yuan, “dscnn: hardware-oriented optimization for stochastic computing based deep convolutional neural networks,” in proc. of the 34th ieee international conference on computer design (iccd), 2016. (ccf b)
26. [socc'16] a. ren, b. yuan, and y. wang, “design of high-speed low-power polar bp decoder using emerging technologies,” in proc. of the 29th ieee international system-on-chip conference (socc), 2016.
27. [isvlsi'16] r. cai, a. ren, y. wang, and b. yuan, “memristor-based discrete fourier transform for improving performance and energy efficiency,” in proc. of ieee computer society annual symposium on vlsi (isvlsi), 2016.
journal papers
28. [jssc'20] j. yue, r. liu, w. sun, z. yuan, z. wang, y. tu, y. chen, a. ren, y. wang, m. chang, x. li, h. yang, and y. liu, “sticker-t: an energy efficient neural network processor using block-circulant algorithm and unified frequency-domain acceleration,” in the ieee journal of solid state circuits (jssc), 2020. (sjr q1)
29. [2d mater. '19] y. qiang, a. ren*, x. zhang, p. patel, x.n han, k. seo, z. shi, y. wang, and h. fang, “design of atomically-thin-body field-effect sensors and pattern recognition neural networks for ultra-sensitive and intelligent trace explosive detection,” in 2d materials, 2019. (sjr q1)
30. [integration'19] j. li, z. yuan, z. li, a. ren, c. ding, j. draper, s. nazarian, q. qiu, b. yuan, and y. wang, “normalization and dropout for stochastic computing-based deep convolutional neural networks,” in integration, the vlsi journal, 2017. (sjr q1)
31. [tcad'18] z. li, j. li, a. ren, r. cai, c. ding, x. qian, j. draper, b. yuan, j. tang, q. qiu, and y. wang, “heif: highly efficient stochastic computing based inference framework for deep neural networks,” in ieee transactions on computer-aided design of integrated circuits and systems, 2018.(ccf a, sjr q2)
32. [nano commun. netw. '18] r. cai, a. ren, s. soundarajan, and y. wang, “a low-computation-complexity, energy-efficient, and high-performance linear program solver based on primal–dual interior point method using memristor crossbars,” in nano communication networks, 2018. (sjr q2)
patent applications
1. a. ren, y. wang, t. zhang, y. xie, “structured pruning for machine learning model,” us patent
2. a. ren, t. zhang, y. wang, y. xie, “method and system for processing a neural network,” us patent
3. f. sun, a. ren, “artificial neural network with sparse weights,” us patent