Ahmed Ali

ORCID: 0000-0002-9186-7544
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Speech Recognition and Synthesis
  • Natural Language Processing Techniques
  • Analog and Mixed-Signal Circuit Design
  • Speech and Audio Processing
  • Music and Audio Processing
  • Topic Modeling
  • Phonetics and Phonology Research
  • Speech and dialogue systems
  • Advancements in Semiconductor Devices and Circuit Design
  • CCD and CMOS Imaging Sensors
  • Radio Frequency Integrated Circuit Design
  • Sensor Technology and Measurement Systems
  • Advancements in PLL and VCO Technologies
  • Misinformation and Its Impacts
  • Multi-Criteria Decision Making
  • Linguistic Variation and Morphology
  • Low-power high-performance VLSI design
  • Authorship Attribution and Profiling
  • Hate Speech and Cyberbullying Detection
  • Language, Linguistics, Cultural Analysis
  • Hydraulic Fracturing and Reservoir Analysis
  • VLSI and Analog Circuit Testing
  • Sentiment Analysis and Opinion Mining
  • Deception detection and forensic psychology
  • Social Media and Politics

Zagazig University
2022-2024

University of Duhok
2020-2024

Cairo University
2003-2023

Qatar Airways (Qatar)
2012-2023

Halliburton (United Kingdom)
2023

Apple (United States)
2022

Hamad bin Khalifa University
2015-2021

Analog Devices (United States)
2005-2020

Suez University
2020

Assiut University
2020

This paper describes a 16-bit 250 MS/s ADC fabricated on 0.18 BiCMOS process. The has an integrated input buffer with new linearization technique that improves its distortion by 5-10 dB and lowers power consumption 70% relative to the state of art. It demonstrates background calibration correct residue amplifier (RA) gain errors lower consumption. summing node sampling (SNS) is based summing-node voltage using it corresponding estimate open loop gain. achieves SNDR 76.5 consumes 850 mW from...

10.1109/jssc.2010.2073194 article EN IEEE Journal of Solid-State Circuits 2010-10-19

We discuss a 14 bit 1 GS/s RF sampling pipelined ADC that utilizes correlation-based background calibration to correct the inter-stage gain, settling and memory errors. To improve linearity performance, employs input distortion cancellation digital technique compensate for non-linear charge injection (kick-back) from capacitors on driver. In addition, an effective dithering is embedded in signal break dependence of calibration's convergence amplitude. The fabricated 65 nm CMOS process has...

10.1109/jssc.2014.2361339 article EN IEEE Journal of Solid-State Circuits 2014-10-21

Marcos Zampieri, Shervin Malmasi, Nikola Ljubešić, Preslav Nakov, Ahmed Ali, Jörg Tiedemann, Yves Scherrer, Noëmi Aepli. Proceedings of the Fourth Workshop on NLP for Similar Languages, Varieties and Dialects (VarDial). 2017.

10.18653/v1/w17-1201 article EN cc-by 2017-01-01

In this paper, we investigate different approaches for dialect identification in Arabic broadcast speech.These methods are based on phonetic and lexical features obtained from a speech recognition system, bottleneck using the i-vector framework.We studied both generative discriminative classifiers, combined these multi-class Support Vector Machine (SVM).We validated our results an Arabic/English language task, with accuracy of 100%.We also evaluated binary classifier to discriminate between...

10.21437/interspeech.2016-1297 article EN Interspeech 2022 2016-08-29

In this paper we present a recipe and language resources for training testing Arabic speech recognition systems using the KALDI toolkit. We built prototype broadcast news system 200 hours GALE data that is publicly available through LDC. describe in detail decisions made building system: MADA toolkit text normalization vowelization; why use 36 phonemes; how generate pronunciations; build model. report results state-of-the-art modeling decoding techniques. The scripts are released on QCRI's...

10.1109/slt.2014.7078629 article EN 2022 IEEE Spoken Language Technology Workshop (SLT) 2014-12-01

We discuss a 12-b 18-GS/s analog-to-digital converter (ADC) implemented in 16-nm FinFET process. The ADC is composed of an integrated high-speed track-and-hold amplifier (THA) driving up to eight interleaved pipeline ADCs that employ open-loop inter-stage amplifiers. Up 10 GS/s, the THA operates at full sampling rate using non-interleaved single sample network, thereby eliminating interleaving time and bandwidth mismatch. Above programmed use two ping-ponged, or optional (2 + 1) randomized,...

10.1109/jssc.2020.3023882 article EN IEEE Journal of Solid-State Circuits 2020-09-30

This paper describes the Arabic MGB-3 Challenge - Speech Recognition in Wild. Unlike last year's MGB-2 Challenge, for which recognition task was based on more than 1,200 hours broadcast TV news recordings from Aljazeera programs, emphasises dialectal using a multi-genre collection of Egyptian YouTube videos. Seven genres were used data collection: comedy, cooking, family/kids, fashion, drama, sports, and science (TEDx). A total 16 videos, split evenly across different genres, divided into...

10.1109/asru.2017.8268952 article EN 2021 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) 2017-12-01

This paper describes a 14-bit, 125 MS/s IF/RF sampling pipelined A/D converter (ADC) that is implemented in 0.35 <tex xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">$muhbox m$</tex> BiCMOS process. The ADC has sample-and-hold circuit integrated the first pipeline stage, which removes need for dedicated amplifier (i.e., "SHA-less"). It also buffer turned off during hold clock phases to save power. To accurately estimate and minimize jitter, new...

10.1109/jssc.2006.875291 article EN IEEE Journal of Solid-State Circuits 2006-07-26

Measuring the performance of automatic speech recognition (ASR) systems requires manually transcribed data in order to compute word error rate (WER), which is often time-consuming and expensive. In this paper, we propose a novel approach estimate WER, or e-WER, does not require gold-standard transcription test set. Our e-WER framework uses comprehensive set features: ASR recognised text, character results complement output, internal decoder features. We report for two features; black-box...

10.18653/v1/p18-2004 article EN cc-by 2018-01-01

This paper describes the Arabic Multi-Genre Broadcast (MGB-2) Challenge for SLT-2016. Unlike last year's English MGB Challenge, which focused on recognition of diverse TV genres, this year, challenge has an emphasis handling diversity in dialect speech. Audio data comes from 19 distinct programmes Aljazeera channel between March 2005 and December 2015. Programmes are split into three groups: conversations, interviews, reports. A total 1,200 hours have been released with lightly supervised...

10.1109/slt.2016.7846277 article EN 2022 IEEE Spoken Language Technology Workshop (SLT) 2016-12-01

Dialect identification (DID) is a special case of general language (LID), but more challenging problem due to the linguistic similarity between dialects. In this paper, we propose an end-to-end DID system and Siamese neural network extract embeddings. We use both acoustic features for task on Arabic dialectal speech dataset: Multi-Genre Broadcast 3 (MGB-3). The was trained using three kinds features: Mel-Frequency Cepstral Coefficients (MFCCs), log Mel-scale Filter Bank energies (FBANK)...

10.21437/odyssey.2018-14 article EN 2018-06-06

Predicting the political bias and factuality of reporting entire news outlets are critical elements media profiling, which is an understudied but increasingly important research direction. The present level proliferation fake, biased, propagandistic content online has made it impossible to fact-check every single suspicious claim, either manually or automatically. Thus, been proposed profile look for those that likely publish fake biased content. This makes possible detect "fake news" moment...

10.18653/v1/2020.acl-main.308 article EN cc-by 2020-01-01

We describe a 14-bit 2.5GS/s non-interleaved pipelined ADC that relies on correlation-based background calibrations to correct the inter-stage gain, settling (dynamic), kick-back and memory errors. A new technique is employed inject large dither signal input non-linear driver, another injected any residual non-linearity in pipeline. In order effect of aging comparators, calibration comparator offsets. The fabricated as dual 28nm CMOS process. An optional interleaved mode provided, where two...

10.1109/vlsic.2016.7573537 article EN 2016-06-01

In this paper, we describe Qatar Computing Research Institute's (QCRI) speech transcription system for the 2016 Dialectal Arabic Multi-Genre Broadcast (MGB-2) challenge. MGB-2 is a controlled evaluation using 1,200 hours audio with lightly supervised Our which was combination of three purely sequence trained recognition systems, achieved lowest WER 14.2% among nine participating teams. Key features our are: acoustic models recently introduced Lattice free Maximum Mutual Information (LF-MMI)...

10.1109/slt.2016.7846279 article EN 2022 IEEE Spoken Language Technology Workshop (SLT) 2016-12-01

Emotion recognition from speech signal based on deep learning is an active research area. Convolutional neural networks (CNNs) may be the dominant method in this In paper, we implement two architectures to address problem. The first architecture attention-based CNN-LSTM-DNN model. novel architecture, convo-lutional layers extract salient features and bi-directional long short-term memory (BLSTM) handle sequential phenomena of signal. This followed by attention layer, which extracts a summary...

10.1109/icassp.2019.8683632 article EN ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019-04-17

This paper describes the fifth edition of Multi-Genre Broadcast Challenge (MGB-5), an evaluation focused on Arabic speech recognition and dialect identification. MGB-5 extends previous MGB-3 challenge in two ways: first it focuses Moroccan recognition; second granularity identification task is increased from 5 classes to 17, by collecting data 17 speaking countries. Both tasks use YouTube recordings provide a multi-genre multi-dialectal wild. transcription used about 13 hours transcribed...

10.1109/asru46091.2019.9003960 article EN 2021 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) 2019-12-01

Recent studies show that Gaussian mixture model (GMM) weights carry less, yet complimentary, information to GMM means for language and dialect recognition. However, state-of-the-art recognition systems usually do not use this information. In research, a non-negative factor analysis (NFA) approach is developed weight decomposition adaptation. This modeling, which conceptually simple computationally inexpensive, suggests new low-dimensional utterance representation method using similar of the...

10.1109/taslp.2014.2319159 article EN IEEE/ACM Transactions on Audio Speech and Language Processing 2014-04-22

High sample rate ADCs with high input bandwidth and low power consumption enable direct RF sampling, more integration, flexibility lower cost for communication, instrumentation other applications. The state of the art interleaved converters enables up to 10GS/s 12-14b resolution [1]-[4]. However, increase rate, number sub-ADCs tends increase, which degrades interleaving spurs due sampling time mismatch, increases capacitance, reduces bandwidth, ADC. Randomization helps alleviate impact...

10.1109/isscc19947.2020.9063011 article EN 2022 IEEE International Solid- State Circuits Conference (ISSCC) 2020-02-01

In this paper, we describe a method to collect dialectal speech from YouTube videos create large-scale Dialect Identification (DID) dataset. Using method, collected Arabic known channels 17 speaking countries in the Middle East and Northern Africa. After refinement process, total of 3,000 hours was available for training DID systems, with an additional 57 development testing. For detailed evaluations, data divided into three sub-categories based on segment duration: short (less than 5s),...

10.1109/icassp40776.2020.9052982 article EN ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2020-04-09

End-to-end neural network systems for automatic speech recognition (ASR) are trained from acoustic features to text transcriptions.In contrast modular ASR systems, which contain separately-trained components modeling, pronunciation lexicon, and language the end-to-end paradigm is both conceptually simpler has potential benefit of training entire system on end task.However, such models more opaque: it not clear how interpret role different parts what information learns during training.In this...

10.21437/interspeech.2019-2599 article EN Interspeech 2022 2019-09-13
Coming Soon ...