In this paper, the authors deal with the problem of speech disfluencies analysis for automatic speech recognition. The origin of speech disfluencies may be of different nature: it may be caused by external influence or by internal failure in the planning of speech act. Failures in the speech act planning may be various, among speech disfluencies one may name such as filled pauses, self-repairs and stipulations. Such disfluencies are an obstacle for automatic processing of speech and its transcriptions. Speech corpora with Rich Transcription (the transcription where such phenomena as sentence boundaries, fillers, disfluencies are marked) are used for studying speech disfluencies. Among such corpora are Czech Broadcast Conversation MDE Transcripts and SWITCHBOARD. It is still unclear what knowledge should be used in speech recognition systems to classify and elicit speech disfluencies. That is why there are no appropriate models of them, which could provide automatic disfluencies processing. Methods for such processing may be distinguished between those dealing with disfluencies by means of acoustic models and by means of combined models (acoustic and language ones). But owing to objective reasons (time and expert expenses) researchers frequently use only acoustic modeling in speech recognition systems. There are a lot of papers describing modeling of speech disfluencies as a part of ASR systems. There is also a group of approaches that are meant for speech recognition accuracy increase by separating disfluencies from speech signal in advance or by means of speech transcriptions. Among possible approaches to deal with these phenomena in ASR systems there are those that allow modeling and eliciting disfluencies as separate verbal and paralinguistic elements, and those that ignore them only distinguishing from useful speech, but not telling one from another. There is an alternative method of processing disfluencies as part of language modeling and modeling of unknown words: speech disfluencies may be treated as Unknown Words class, and then building a language model with the account of these phenomena. For the Russian language there are no methods developed for speech disfluencies processing, so it is worth trying to apply different methods and compare results. Due to high expenses of making a corpus of transcripts, which would account for speech disfluencies and would be suitable for training language model (at least 3-gram model), speech disfluencies processing with parametric methods seems to be optimal.