Search
Search Funnelback University
- Refined by:
- Date: 2000
1 -
15 of
15
search results for `broadcast news texts` |u:mi.eng.cam.ac.uk
Fully-matching results
-
The Cambridge University Spoken Document Retrieval System
mi.eng.cam.ac.uk/reports/full_html/johnson_icassp99.html/8 Mar 2000: 132 million words of broadcast news texts, the LDC-distributed 1995 newswire texts, and the transcriptions of the acoustic training data. ... Furthermore new language models were estimated from an extended set of broadcast news transcripts and newspaper -
The 1998 HTK Broadcast News Transcription System: Development and…
mi.eng.cam.ac.uk/reports/full_html/woodland_darpa99.html/2 Mar 2000: The 1997 system used N-gram language models trained on 132 million words of broadcast news texts, the LDC-distributed 1995 newswire texts, and the transcriptions from BNtrain97 (LMtrain97). ... The word N-grams were trained by interpolating (and merging) -
LARGE SCALE MMIE TRAINING FOR CONVERSATIONAL TELEPHONE SPEECH…
mi.eng.cam.ac.uk/reports/full_html/woodland_stw00.html/5 Oct 2000: The system uses word-based N-gram LMs estimated from an interpolation of Hub5 acoustic training transcriptions and Broadcast News texts. -
General Query Expansion Techniques for Spoken Document Retrieval
mi.eng.cam.ac.uk/reports/full_html/jourlin_esca99.html/2 Mar 2000: Our retrieval system was run on 7 different sets of automatically transcribed broadcast news texts with a WER varying from 24.8% to 66.2%. ... J. Young. Segment Generation and Clustering in the HTK Broadcast News Transcription System [in Proc. -
The 1997 HTK Broadcast News Transcription System
mi.eng.cam.ac.uk/reports/full_html/woodland_darpa98.html/1 Mar 2000: The system uses a language model trained on 132 million words of broadcast news texts, the LDC-distributed 1995 newswire texts, and the transcriptions from BNtrain96. ... using the broadcast news training texts, the acoustic training data and 1995 -
Experiments in Broadcast News Transcription
mi.eng.cam.ac.uk/reports/full_html/woodland_icassp98.html/1 Mar 2000: Experiments in Broadcast News Transcription. P.C. Woodland, T. Hain, S.E. Johnson, T. ... Cross-word context dependent decision tree state clustered mixture Gaussian HMMs are used with a 65k word vocabulary and a language model trained on 132 million -
The 1998 HTK Broadcast News Transcription System:Development and…
mi.eng.cam.ac.uk/reports/svr-ftp/woodland_darpa99.pdf8 Mar 2000: The1997 system used N-gram language models trained on 132million words of broadcast news texts, the LDC-distributed1995 newswire texts, and the transcriptions from BNtrain97(LMtrain97). ... 5.1. Language Models. The word N-grams were trained by -
THE CAMBRIDGE UNIVERSITY SPOKEN DOCUMENT RETRIEVAL SYSTEM S.E.…
mi.eng.cam.ac.uk/reports/svr-ftp/johnson_icassp99.pdf8 Mar 2000: The paper is organised as follows. Firstly an overview ofthe HTK broadcast news transcription system used to generatehigh quality automatic transcriptions is given. ... of broadcast news texts, the LDC-distributed 1995newswire texts, and the -
EXPERIMENTS IN BROADCAST NEWS TRANSCRIPTION P.C. Woodland, T. Hain,…
mi.eng.cam.ac.uk/reports/svr-ftp/woodland_icassp98.pdf10 Apr 2000: Cross-word context dependent deci-sion tree state clustered mixture Gaussian HMMs are used with a65k word vocabulary and a language model trained on 132 millionwords of broadcast news texts, along ... Young S.J. (1997) TheDevelopment of the 1996 -
GENERAL QUERY EXPANSION TECHNIQUESFOR SPOKEN DOCUMENT RETRIEVAL…
mi.eng.cam.ac.uk/reports/svr-ftp/jourlin_esca99.pdf10 Apr 2000: different sets ofautomatically transcribed broadcast news texts with aWER varying from24.8% to 66.2%. ... The input data is presented to the system as com-plete episodes of broadcast news shows and these are. -
Spoken Document Retrieval for TREC-8 at Cambridge University
mi.eng.cam.ac.uk/reports/full_html/johnson_trec8.html/30 Mar 2000: Three fixed backoff word-based language models were trained, from broadcast news text, newspaper texts and acoustic transcriptions, which were all generated using data from before 31st January 1998. ... 1996 to Jan. 1998 (from the Primary Source Media -
THE DEVELOPMENT OF THE1996 HTK BROADCAST NEWS TRANSCRIPTION SYSTEM ...
mi.eng.cam.ac.uk/reports/svr-ftp/woodland_darpa97.pdf8 Mar 2000: 4.6. Static Language Models. The evaluation system language model (LM) had a word listcontaining 65423 words chosen from the most frequent wordsin the broadcast news training texts, with the most ... 5. CONCLUSIONThis paper has described our initial -
The 1997 HTK Broadcast News Transcription SystemP.C. Woodland, T. ...
mi.eng.cam.ac.uk/reports/svr-ftp/woodland_darpa98.pdf8 Mar 2000: The 1997 HTK Broadcast News Transcription SystemP.C. Woodland, T. Hain, S.E. ... These were trained on the 132million words of the LDC broadcast news training texts,the transcriptions of the BNtrain97 data (added twice), the1995 newswire texts (both -
LARGE SCALE DISCRIMINATIVE TRAINING FORSPEECH RECOGNITION P.C.…
mi.eng.cam.ac.uk/reports/svr-ftp/woodland_asr00.pdf6 Nov 2000: It was reported in [32] that FD didn’t improveerror rates over MLE trained models for a broadcast newsrecognition task. ... The system uses word-based N-gramLMs estimated from an interpolation of Hub5 acoustic train-ing transcriptions and Broadcast -
SPOKEN DOCUMENT RETRIEVAL FOR TREC-8 AT CAMBRIDGE UNIVERSITY S.E. ...
mi.eng.cam.ac.uk/reports/svr-ftp/johnson_trec8.pdf10 Apr 2000: Three fixed backoff word-based language models were trained,from broadcast news text, newspaper texts and acoustic tran-scriptions, which were all generated using data from before 31stJanuary 1998. ... The first model was built using 190 million wordsof
Search history
Recently clicked results
Recently clicked results
Your click history is empty.
Recent searches
- download (220) · moments ago
Recent searches
Your search history is empty.