Session | Room | Chair | |
Overview Session 1 | Meeting Room 1 | ||
Date | Time | Title | Speaker |
4-Dec | 16:20-16:40 | A Decade of Progress in Sound Event Localization and Detection: Transforming Environmental Sound Analysis for Real-World Impact | Woon-Seng Gan, Nanyang Technological University |
16:40-17:00 | Exploring the Forward-Forward Algorithm: A Novel Learning Approach | Waleed H. Abdulla, The University of Auckland | |
17:00-17:20 | Eye-gaze-based Human-Intention Detection | Kosin Chamnongthai, King Mongkut's University of Technology Thonburi | |
17:20-17:40 | From GPT Evolution to Enterprise Deployment: Key Trends in Generative AI | Jing-Ming Guo, National Taiwan University of Science and Technology | |
17:40-18:00 | An Overview of Online Distributed Kernel Methods for Supervised and Unsupervised Learning | Anthony Kuh, University of Hawaii |
Session | Room | Chair | |
Overview Session 2 | Meeting Room 8 | ||
Date | Time | Title | Speaker |
5-Dec | 10:20-10:40 | An AI-based Diagnostic-aid for Epileptic Electroencephalography | Toshihisa Tanaka, Tokyo University of Agriculture and Technology |
10:40-11:00 | Machine Learning for Analytics Architecture: AI to Design AI Video | Chris Gwo Giun Lee, National Cheng Kung University | |
11:00-11:20 | Compression of Large AI Models | Weisi Lin, Nanyang Technological University | |
11:20-11:40 | Introduction to Multi-Camera Systems and 3D Quality Assessment | Sanghoon Lee, Yonsei University | |
11:40-12:00 | Highlight of New Image Generative Models and Applications to Image Manipulations | Wan-Chi Siu, Hong Kong Polytechnic University & St. Francis University |
Session | Room | Chair | |
Overview Session 3 | Merged Room (Room 10 + 11) | ||
Date | Time | Title | Speaker |
6-Dec | 9:00-9:20 | Overview of Source Camera Identification Techniques | Bonnie N. F. Law, The Hong Kong Polytechnic University |
9:20-9:40 | Recent Advances in Complete Quality Preserving Data Hiding | KokSheik Wong, Monash University Malaysia | |
9:40-10:00 | Real or Fake? Frontiers of Countering Fake Media in the Age of Infodemics | Isao Echizen, National Institute of Informatics | |
10:00-10:20 | User Preference Modeling and Analysis in Choice Problems | H. Vicky Zhao, Tsinghua University |
In this talk, I will outline the historical progression of computational Auditory Scene Analysis, with a particular emphasis on the sound event detection and localization (SELD) task. SELD aims to enhance the human auditory system by automating extended environmental listening, with applications spanning noise monitoring, surveillance, real-sound control for mobile devices, and biodiversity tracking. Over the past decade, substantial progress has been made in SELD, driven in large part by the Detection and Classification of Acoustic Scenes and Events (DCASE) challenge series initiated since 2010. This presentation will spotlight major breakthroughs in SELD, examine recent advancements and emerging research directions, and explore real-world applications. Additionally, I will share some of the latest research findings and proof-of-concept prototypes developed by my research group.
WOON-SENG GAN is a Professor of Audio Engineering and Director of the Smart Nation TRANS
(national) Lab in the School of Electrical and Electronic Engineering at Nanyang
Technological University, Singapore. He received his BEng (1st Class Hons) and Ph.D.
degrees, both in Electrical and Electronic Engineering from the University of
Strathclyde, the UK in 1989 and 1993, respectively. He has held several leadership
positions at Nanyang Technological University, including serving as the Head of the
Information Engineering Division from 2011 to 2014, and as the Director of the Centre
for Info-comm Technology from 20016 to 20019. Through his involvement in the Smart
Nation TRANS lab, he led the university effort in translating AI research into
real-world deployment.
Prof. Gan is a highly distinguished researcher and academic with numerous prestigious
affiliations and editorial roles. His expertise and contributions have earned him
fellowships in esteemed organizations, including the Audio Engineering Society (AES) and
the Institute of Engineering and Technology (IET), and the Institute of Electrical and
Electronics Engineers (IEEE), where he holds the status of a Senior Member. He has
served as an Associate Editor for the IEEE/ACM Transaction on Audio, Speech, and
Language Processing (TASLP) from 2012 to 2015. His dedication to the journal's
excellence was rewarded with the Outstanding IEEE TASLP Editorial Board Service Award in
2016, a testament to his contributions and leadership in scholarly publishing.
Currently, Prof. Gan holds significant editorial responsibilities as a Senior Area
Editor for the IEEE Signal Processing Letters(2019-2024), contributing to the
publication's high-quality content and its impact on the signal processing community.
Additionally, he serves as an Associate Technical Editor for the Journal of Audio
Engineering Society (JAES)and Associate Editor of the prestigious IEEE Signal Processing
Magazine. He has been featured among the "World's Top 2% Scientist 2021" by Stanford
University, solidifying his place as one of the most influential researchers in the
field of Acoustics. He is recently elected as President for the Asia-Pacific Signal and
Information Processing Association and starting his tenure from 2025.
The Forward-Forward (FF) Algorithm is a recent learning procedure for neural networks based on
Boltzmann machines and Noise Contrastive Estimation. This algorithm eliminates both the
backpropagation's forward and backward passes and substitutes them with two forward passes, each
on different data but with different goals. The positive pass employs actual data to raise
“goodness” in the hidden layers, and the negative pass employs “negative data” to lower it.
The FF algorithm is especially useful when specifics of forward computation are unclear so that
learning can take place without preserving neural activities or passing error derivatives. It
takes nearly the same prediction time to perform as backpropagation and can be applied where
reinforcement learning would normally be needed. While it may not be as good as backpropagation
for specific problem types, it has prospects for use in low-power analog VLSI and as a model for
learning in the cortex.
In this presentation, I will present some of the initial findings of the FF algorithm, apply it
to small networks, and establish a basis for its application to larger networks. We will also
talk about various measures of “goodness” and the general conclusion about the algorithm’s
impact on further analysis of neural networks.
WALEED H. ABDULLA is an Associate Professor at the University of Auckland, where he has
been a faculty member since 2002. He holds a Ph.D. in signal processing from the
University of Otago, New Zealand. A founding member of APSIPA, Waleed has served as Vice
President for Member Relations and Development, the first Editor in-Chief of the APSIPA
Newsletter, and is currently a member of its Board of Governors. He is also an APSIPA
Distinguished Lecturer. Waleed has collaborated with leading institutions worldwide and
supervised over 35 Ph.D. students. Recognized for his teaching excellence, he has
received multiple awards, including the Otago University Scholarship and teaching awards
in 2005 and 2012. His research focuses on signal processing classical and deep learning
techniques, with applications in biometrics, speech and speaker recognition,
hyperspectral imaging, noise control, and audio watermarking. He is the author of a
notable book on audio watermarking and has contributed over 300 publications with 3000+
citations. As a senior member of IEEE and a life member of APSIPA, Waleed has also
delivered well-received tutorials at APSIPA conferences.
Recently, many robots have been working with humans in many places such as manufacturing factories, hospitals, department stores, airports, construction sites, and so on. Sometimes, accidents occur by robot and human crashes, since the robot has no function to sense human intention, while a man has some senses to estimate the intention of another man by detecting eye gaze. Eye gaze detection becomes an important basic function, which would be extended to detect intention and other applications such as intention reading for stroke patients, customer intention estimation, worker intention estimation for collaborative robots, game entertainment, and so on. This talk will update the technologies related to eye-gaze-based human-intention detection and discuss the trend in future research.
KOSIN CHAMNONGTHAI (S'85-M'87-SM'91) received a B.Eng. degree in applied
electronics from the University of Electro-Communications in 1985, an M.Eng.
degree in electrical engineering from Nippon Institute of Technology in 1987,
and a Ph.D. degree in electrical engineering from Keio University in 1991. He is
currently a Professor with the Electronic and Telecommunication Engineering
Department, Faculty of Engineering, King Mongkut's University of Technology
Thonburi. He served as vice president (conference) of APSIPA Association
(2020-2023), Editor of ECTI e-magazine from 2011 to 2015, an Associate Editor of
ECTI-EEC Transactions, from 2003 to 2010, and ECTI-CIT Transactions, from 2011
to 2016. He has served as the Chairman of the IEEE COMSOC Thailand from 2004 to
2007, and the President of the ECTI Association (2018-2019). His research
interests include computer vision, image processing, robot vision, signal
processing, and pattern recognition. He is a senior member of IEEE, and a member
of IEICE, TESA, ECTI, AIAT, APSIPA, TRS, and EEAAT.
This presentation explores the latest trends in Generative AI, focusing on key topics such as the evolution of GPT, domain-specific optimization in large language models (LLMs), micro-sizing LLMs, and the deployment of Generative AI in Taiwan's Industrial Technology Research Institute (ITRI) and its potential expansion to Taiwanese enterprises. The talk delves into the advancements in GPT models, highlights the significance of expert-level optimization techniques in enhancing LLM performance, and discusses strategies for scaling down large models. Moreover, it showcases the current efforts by ITRI in implementing Generative AI and outlines the future prospects of adopting this technology within Taiwan's business landscape.
JING-MING GUO is currently a full Professor with the Department of Electrical
Engineering, and Director of Advanced Intelligent Image and Vision Technology Research
Center. He was Vice Dean of the College of Electrical Engineering and Computer Science,
and Director of the Innovative Business Incubation Center, Office of Research and
Development. His research interests include Big data signal processing, artificial
intelligence, digital image/video processing, computer vision. Dr. Guo is a Senior
Member of the IEEE and Fellow of the IET.
Dr. Guo is Chapter Chair of IEEE Signal Processing Society, Taipei Section, Board of
Governor member of Asia Pacific Signal and Information Processing Association, and
President of IET Taipei Local Network. He will be/was General Chair of many
international conferences, e.g., APSIPA 2023, IEEE Life Science Workshop 2020, ISPACS
2019, IEEE ICCE-Berlin 2019, IWAIT 2018, and IEEE ICCE-TW 2015. He will be/was Technical
Program Chair of many international conferences as well, e.g., IEEE ICIP 2023, IWAIT
2022, IEEE ICCE-TW 2014, IEEE ISCE 2013, and ISPACS 2012. He is/was Associate Editor of
the IEEE Transactions on Image Processing, IEEE Transactions on Circuits and Systems for
Video Technologies, IEEE Transactions on Multimedia, IEEE Signal Processing Letters,
Information Sciences, Signal Processing, and Journal of Information Science and
Engineering.
This talk discusses nonlinear filtering algorithms using kernel methods. We start by discussing the problem for centralized online kernel methods using Random Fourier Features. We then discuss federated online learning algorithms discussing tradeoffs between performance, computational complexity, and communication costs. Finally, we present some examples for time series prediction and anomaly detection.
ANTHONY KUH received his B.S. in Electrical Engineering and Computer Science at the
University of California, Berkeley in 1979, an M.S. in Electrical Engineering from
Stanford University in 1980, and a Ph.D. in Electrical Engineering from Princeton
University in 1987. He previously worked at AT&T Bell Laboratories and has been on the
faculty in the Department of Electrical and Computer Engineering at the University of
Hawai’i since 1986. He is currently a Professor and previously served as Department
Chair. His research is in the area of neural networks and machine learning, adaptive
signal processing, sensor networks, and renewable energy and smart grid applications. He
won a National Science Foundation (NSF) Presidential Young Investigator Award and is an
IEEE Life Fellow. He is currently serving as a program director for NSF in the
Electrical, Communications, and Cyber Systems (ECCS) division working in the Energy,
Power, Control, and Network (EPCN) group. He previously served for the IEEE Signal
Processing Society on the Board of Governors as a Regional Director-at-Large Regions1-6,
as a senior editor for the IEEE Journal of Selected Topics in Signal Processing, and as
a member of the Awards Board. He currently serves on the Board of Governors as the
Immediate Past President of the Asia Pacific Signal and Information Processing
Association (APSIPA).
This talk addresses an "AI-based diagnosis and treatment-aid" system that learns from electroencephalograms (EEG) of epilepsy patients and diagnoses made by clinicians. Japan has about one million epilepsy patients, but there is a very limited number (about 800) of clinicians who can properly read and interpret patient EEG. Epileptic seizures sometimes cause serious traffic accidents, highlighting the need for effective social measures to address this issue. We have developed AI-based automated algorithms that can learn from diagnoses made by medical specialists in epilepsy for EEG data measured in hospitals. I will report techniques to construct a dataset of EEG and to learn the interpretation of the data.
TOSHIHISA TANAKA received the B.E., M.E., and Ph.D. degrees from the Tokyo Institute of
Technology, in 1997, 2000, and 2002, respectively. From 2000 to 2002, he was a JSPS
Research Fellow. From October 2002 to March 2004, he was a Research Scientist with the
RIKEN Brain Science Institute. In April 2004, he joined the Department of Electrical and
Electronic Engineering, Tokyo University of Agriculture and Technology, where he is
currently a Professor. In 2005, he was a Royal Society Visiting Fellow with the
Communications and Signal Processing Group, Imperial College London, U.K. From June 2011
to October 2011, he was a Visiting Faculty Member of the Department of Electrical
Engineering, University of Hawaii at Manoa. His research interests include signal
processing and machine learning, including brain and biomedical signal processing,
brain–machine interfaces, and adaptive systems. Prof. Tanaka is a member of IEICE,
APSIPA, the Society for Neuroscience, and the Japan Epilepsy Society. He is the
Co-Founder and the CTO of Sigron, Inc. Furthermore, he served as a Member-at-Large on
the Board of Governors (BoG) of the Asia–Pacific Signal and Information Processing
Association (APSIPA). He was a Distinguished Lecturer of APSIPA. He serves as a
Vice-President of APSIPA. He served as an Associate Editor and a Guest Editor for
special issues in journals, including IEEE ACCESS, Neurocomputing, IEICE Transactions on
Fundamentals, Computational Intelligence and Neuroscience (Hindawi), IEEE TRANSACTIONS
ON NEURAL NETWORKS AND LEARNING SYSTEMS, Applied Sciences (MDPI), and Advances in Data
Science and Adaptive Analysis (World Scientific). He served as the Editor-in-Chief of
Signals (MDPI). Currently, he serves as an Associate Edi tor for Neural Networks
(Elsevier). He is a Co-Editor of Signal Processing Techniques for Knowledge Extraction
and Information Fusion (with Mandic, Springer, 2008), and a leading Co-Editor of Signal
Processing and Machine Learning for Brain-Machine Interfaces (with Arvaneh, IET, U.K.,
2018).
Niklaus Emil Wirth introduced the innovative idea that Programming = Algorithm + Data Structure.
Inspired by this, we advance the concept to the next level by stating that Design = Algorithm +
Architecture. With concurrent exploration of algorithm and architecture entitled
Algorithm/Architecture Co-exploration (AAC), this methodology introduces a leading paradigm
shift in advanced system design from System-on-a-Chip to Cloud and Edge.
As algorithms with high accuracy become exceedingly more complex and Edge/IoT generated data
becomes increasingly bigger, flexible parallel/reconfigurable processing are crucial in the
design of efficient signal processing systems having low power. Hence the analysis of algorithms
and data for potential computing in parallel, efficient data storage and data transfer is
crucial. With extension of AAC for SoC system designs to even more versatile platforms based on
analytics architecture, system scope is readily extensible to cognitive cloud and reconfigurable
edge computing for multimedia and mobile health, a cross-level-of abstraction topic which will
be introduced in this tutorial together with case studies in lightweight mobile edge for skin
cancer detection with two layers CNN at 97% recognition rate. High level Synthesis (HLS) may
also be presented should time allows.
CHRIS GWO GIUN LEE is an investigator in signal processing systems for multimedia and
bioinformatics. His work on analytics of algorithm concurrently with architecture,
Algorithm/Architecture Co-Design (AAC), has made possible accurate and efficient
computations on SoC, cloud and edge including Digital Health. He is currently using AI
to Design AI and is also enabling accessible health and wellness via AI Humanity.
Chris’ work has contributed to 130+ original research and technical publications
with the invention of 50+ patents worldwide. His AAC work was used by the industry in
deploying more than 60 million LCD panels worldwide. Two of these patents were also
licensed by US health industry for development of analytics platform based precision
medicine products (Boston, MA, June 1, 2015, GLOBE NEWSWIRE). Chris’ AAC work has also
been pivotal in delivering feasible and realistic international standards, including 3D
extension of HEVC and Reconfigurable Video Coding in ISO/IEC/MPEG, for applications
requiring processing of big multimedia data. His low-complexity 3D video coding
technology was also included in MPEG.
Chris worked for Philips Semiconductor as a system architect and project leader in
the Silicon Valley. He was recruited to National Cheng Kung University in 2003. He has
been conducting multidisciplinary research having collaborations with: IBM TJ Watson
Research Center on Cloud/Reconfigurable Computing; Banner Alzheimer Institute on
Intelligent Health; National Center for High-performance Computing, National Research
Laboratory, Taiwan on medical image Analytics; National Taiwan University on
Harmonically Generated Microscopy Medical Image Processing; and National Cheng Kung
University Medical Center, Kaohsiung Medical University Center on telehealth. He has
been working with Google on Silicon AI and is working with Cadence on High Level
Synthesis.
Chris received his B.S. degree in electrical engineering from National Taiwan
University, and M.S. and Ph.D. degrees in electrical engineering from the University of
Massachusetts. He served as the AE for IEEE TSP (2016 ~ 2020) and Journal of Signal
Processing Systems from 2010 till now. He was formerly the AE for IEEE TCSVT (2009 ~
2014) for which he received the Best Associate Editor’s Award in 2011. Chris is the IEEE
CASS Distinguished Lecturer from 2019 to 2021. Having served as ExCom Member for IEEE
Region 10 starting 2017, he also chaired the Industry Relations Committee from 2019 to
2020. Chris will chair Region 10’s Award Recognition starting 2025.
Chris served as APSIPA’s VP of IRD and was the Deputy VP of IRD from 2020 to 2021,
Special Session Co-Chair, APSIPA 2018, TPC Co-Chair, APSIPA 2017, and Chair for APSIPA
SPS-TC from 2015-2016 with establishment of sub-committees. He also established APSIPA
Taiwan Chapter since May 2020. Chris is the founder and CTO for CogniNU Technologies on
Digital Health.
With the unprecedented effort by academy and industry recently, large AI models have demonstrated impressive achievement and great potential in diverse areas; however, such models demand rapidly increasing resources in energy, computation and storage. To make them green and economical is the key toward sustainable, responsible and affordable solutions. In this talk, we will discuss into issues and opportunities for pruning and compressing large deep neural networks, in order to improve implementational efficiency, hardware design, power consumption, and so on, without jeopardizing system performance.
WEISI LIN received the Ph.D. degree from King’s College London, U.K. He is currently a
Professor with the College of Computing and Data Science, Nanyang Technological
University. His research interests include image processing, perceptual signal modeling,
video compression, and multimedia communication, in which he has published over 200
journal articles, over 230 conference papers, filed seven patents, and authored two
books. He is a fellow of IET and an Honorary Fellow of Singapore Institute of
Engineering Technologists. He was the Technical Program Chair of the IEEE ICME 2013, PCM
2012, QoMEX 2014, and IEEE VCIP 2017. He has been an invited/panelist/keynote/tutorial
speaker at over 20 international conferences. He has been an Associate Editor of the
IEEE TRANSACTIONS ON IMAGE PROCESSING, the IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR
VIDEO TECHNOLOGY, the IEEE TRANSACTIONS ON MULTIMEDIA, and the IEEE SIGNAL PROCESSING
LETTERS. He was a Distinguished Lecturer of the IEEE Circuits and Systems Society from
2016 to 2017 and the Asia–Pacific Signal and Information Processing Association (APSIPA)
from 2012 to 2013.
As the boundaries between Computer Vision and Graphics continue to blur, the use of 3D Human Avatars through Vision technology is gaining increasing attention. With these technological advancements, the ability to capture high-quality 3D data has opened up new possibilities for creating more refined 3D Avatars using Deep Learning. Setting up a camera system plays a key role in this process, offering a more efficient and essential method for acquiring 3D data. In this presentation, I will introduce the camera system setup, discuss 3D Human Avatars, and explore the technologies used to evaluate them. Defining mesh quality features is challenging due to the irregular topology of meshes, which are defined on vertices and triangles. To address this, we propose a novel 3D projective structural similarity index (3D-PSSIM) for meshes that is robust to differences in mesh topology. Additionally, to evaluate the quality of the generated human motion, we introduce various metrics for human pose and motion analysis.
SANGHOON-LEE received the B.S. degree from Yonsei University, South Korea, in 1989, the
M.S. degree from the KAIST, South Korea, in 1991, and the Ph.D. degree from The
University of Texas at Austin, Austin, TX, USA, in 2000. From 1991 to 1996, he was with
Korea Telecom, South Korea. From 1999 to 2002, he was with Lucent Technologies, NJ, USA.
In 2003, he joined the Department of Electrical and Electronic Engineering, Yonsei
University, as a Faculty Member, where he is currently a Full Professor. He was a member
of the IEEE IVMSP/MMSP TC from 2014 to 2019 and from 2016 to 2021, respectively. He is a
BoG Member of APSIPA. He was the General Chair of the 2013 IEEE IVMSP Workshop. He has
been serving as the Chair for the IEEE P3333.1 Working Group since 2011. He was an
Associate Editor and a Guest Editor of the IEEE Transactions on Image Processing from
2010 to 2014, and 2013, respectively. He served as an Associate Editor for the IEEE
Signal Processing Letters from 2014 to 2018. He has been serving as a Senior Area Editor
for the IEEE Signal Processing Letters since 2018 and an Associate Editor and Deputy
Editor-in-Chief for the IEEE Transactions on Multimedia 2022 and 2024, respectively.
In this overview, we are going to highlight the essences some recent generative models and give an introduction to our new domain transfer generative model. We subsequently will discuss briefly the ways to make various applications, such as face generation, image inpainting, image/video super-resolution, face manipulations and video coding.
WAN-CHI SIU is Emeritus Professor (former Chair Professor, HoD(EIE) and Dean(FEng)) of
Hong Kong Polytechnic University, and Research Professor of St. Francis University. He
chaired the First Engineering/IT Panel of the Research Assessment Exercise in Hong Kong
(RAE, 1992/3). He was Vice President, and Chair of Conference Board of the IEEE SP
Society (2012-2014), and President of APSIPA (2017-2018). He is Life-Fellow of IEEE, and
has been Guest Editor/Subject Editor/AE for IEEE Transactions on CAS, IP, CSVT, and
Electronics Letters, published over 500 research papers in signal processing, fast
algorithms, AI, deep learning, video coding, computer vision and pattern recognition. He
organized many international conferences, including IEEE society-sponsored flagship
conferences, as TPC Chair (ISCAS1997) and General Chair (ICASSP2003 and ICIP2010).
One of the key problems in image forensics is the source camera identification which aims to
uncover the origins of digital photos. Traditionally, signal processing (SP) techniques
constructed camera signatures to uniquely characterize capturing devices. With the surge of deep
learning methodologies across diverse research fields, these techniques have been used for
source camera identification tasks. Architectures such as convolutional neural networks (CNNs)
and Residual Networks (ResNets) have been explored. However, studies found that these approaches
are effective up to the model level, i.e., the camera model used to capture the photo can be
identified. Accurately identifying the specific camera device remains a challenge compared to
traditional SP methods.
SP-based and DL approaches require either the source camera to produce a camera fingerprint
or substantial photos captured by the device for model training. This requirement posts hurdles
in practical real-world scenario. To address this limitation, the source camera identification
problem can be reformulated as an image similarity matching problem in which contrastive
learning can be used. It offers a promising solution to enhance performance and address
practical constraints. In this talk, an overview of source camera identification is given. We
will highlight the principles and results of different methods, including traditional SP
approaches, deep learning approaches and the recent contrastive learning approaches. Various
challenges will also be discussed.
BONNIE LAW is currently an associate professor at the Hong Kong Polytechnic University.
Her research interests focus on signal and image processing, in particular image
forensics which involves works on seam carving and source camera identification using
deep learning. Throughout the years, she has engaged in various professional activities.
For example, she served as the general co-chair of the 2017 International Conference on
SP, Communication and Computing. She also held positions as the Treasurer of the IEEE HK
Chapter of SP from 2009 to 2016 and as the secretary of APSIPA Headquarters since 2009.
Data hiding aims to insert data into content to achieve specific purposes, including claiming
ownership through watermarking, validating the integrity of the content in authentication, and
providing extra/paid features, to name a few. Data hiding concerns three conflicting requirement
aspects, namely, quality, capacity, and robustness. For many years, researchers have aimed to
insert as much data as possible while minimizing quality deterioration and maximizing
robustness. However, in a typical data hiding method where the content is modified to encode
data, the quality of the host is always degraded.
In this overview, the concept of complete quality preservation is introduced, and the need
for such feature is discussed. Subsequently, recent techniques designed to hide data into PDF
file and animated image while completely preserving their qualities are introduced, followed by
some potential research directions and use cases.
KOKSHEIK WONG is an Associate Professor in the School of Information Technology at
Monash University Malaysia. In 2009, he received the Doctor of Engineering degree from
Shinshu University, Japan, under the Monbukagakusho scholarship.
KokSheik's research interests include multimedia signal processing and
cybersecurity. He actively contributes to the research community as associate editor at
IEEE Signal Processing Letters, and the Journal of Information Security and
Applications. He is a member of the Information Forensics and Security (IFS) technical
committee of the IEEE Signal Processing Society, where he takes up an additional role as
an IFS representative in the Challenges and Data Collections committee. In addition, he
is a member of the Board of Governors: Members-at-large in the Asia Pacific Signal and
Information Processing Association (APSIPA), and a member of the APSIPA Multimedia
Security and Forensics (MSF) technical committee.
Recently, Dr. Wong has ventured into digital health, where he leads a work package
in a European Union grant - WAge (101137207) to research on evidence-based interventions
for promotion of mental and physical health in changing working environments
(post-pandemic workplaces).
AI can now learn from human-derived information such as the face, voice, body, and natural language to generate synthetic media that can be mistaken for the real thing. Synthetic media are used in a variety of applications in the fields of communication and entertainment. On the other hand, as a negative aspect of synthetic media, there are cases in which fake media, such as fake video, fake audio, and fake text, are generated and distributed for fraud, inducement of thought, and manipulation of public opinion. There is a fear of an "infodemic," a flood of uncertain information that causes anxiety and confusion in society. This talk will give an overview of the threats posed by such fake media and introduce the latest results of the research project "Social Information Technologies to Counter Infodemics" (CREST FakeMedia), which the speakers are currently working on under the Strategic Basic Research Program (CREST) of the Japan Science and Technology Agency (JST). Ref: CREST FakeMedia, http://research.nii.ac.jp/~iechizen/crest-e.html
ISAO ECHIZEN received B.S., M.S., and D.E. degrees from the Tokyo Institute of
Technology, Japan, in 1995, 1997, and 2003, respectively. He joined Hitachi, Ltd. in
1997 and until 2007 was a research engineer in the company's systems development
laboratory. He is currently a director and a professor of the Information and Society
Research Division, the National Institute of Informatics (NII), a director of the Global
Research Center for Synthetic Media, the NII, and a professor in the Department of
Information and Communication Engineering, Graduate School of Information Science and
Technology, the University of Tokyo, Japan. He is currently engaged in research on
multimedia security and multimedia forensics. He was a member of the Information
Forensics and Security Technical Committee of the IEEE Signal Processing Society. He is
the IEICE Fellow, the IPSJ Fellow, the IEEE Senior Member, and the Japanese
representative on IFIP and on IFIP TC11 (Security and Privacy Protection in Information
Processing Systems), a member-at-large of the board-of-governors of APSIPA, and an
editorial board member of the IEEE Transactions on Dependable and Secure Computing, the
EURASIP Journal on Image and Video processing, and the Journal of Information Security
and Applications, Elsevier.
We make decisions from many choices with conflicting attributes every day. Although these decision-making problems are diverse, a user's decision is often influenced by three factors: the inter-competition competition, that is, the competition among available choices; the context effect where a user’s preference for a seller is affected by the market where the seller is in; and the projection bias where users are biased when estimating their future preference. A challenging issue here is to model a user's personal preference and to understand his/her decision process, and this is especially important in the information era, where users are often overwhelmed by the avalanche of information available online. In this talk, we will talk about our recent works on model-based methods to analyze the impact of the above factors on users’ choices in e-commerce. This investigation can help address the information overload problem, offer personalized services, and provide essential guidelines on seller pricing strategies, market demand analysis, etc.
H. VICKY ZHAO received her B.S. and M.S. degree from Tsinghua University, China, in 1997
and 1999, respectively, and her Ph. D degree from the University of Maryland, College
Park, in 2004, all in electrical engineering. She was with the Department of Electrical
and Computer Engineering at the University of Alberta, Canada, from 2006 to 2016. She is
currently an Associate Professor in the Department of Automation at Tsinghua University,
Beijing, China. Her research interests include social network analysis, information
security and forensics, digital signal processing and communications. Dr. Zhao received
the IEEE Signal Processing Society (SPS) 2008 Young Author Best Paper Award and the 2020
APSIPA Annual Summit and Conference Best Paper Award. She has co-authored 3 research
monographs. She serves as a Member-at-Large of the IEEE Signal Processing Society Board
of Governors (2022-2024), and a Member-at-Large of the APSIPA Board of Governors (2024).
She is an APSIPA Distinguished Lecturer (2023-2024).