Resumen: This paper presents a new approach based on recurrent neural networks (RNN) to the multiclass audio segmentation task whose goal is to classify an audio signal as speech, music, noise or a combination of these. The proposed system is based on the use of bidirectional long short-term Memory (BLSTM) networks to model temporal dependencies in the signal. The RNN is complemented by a resegmentation module, gaining long term stability by means of the tied state concept in hidden Markov models. We explore different neural architectures introducing temporal pooling layers to reduce the neural network output sampling rate. Our findings show that removing redundant temporal information is beneficial for the segmentation system showing a relative improvement close to 5%. Furthermore, this solution does not increase the number of parameters of the model and reduces the number of operations per second, allowing our system to achieve a real-time factor below 0.04 if running on CPU and below 0.03 if running on GPU. This new architecture combined with a data-agnostic data augmentation technique called mixup allows our system to achieve competitive results in both the Albayzín 2010 and 2012 evaluation datasets, presenting a relative improvement of 19.72% and 5.35% compared to the best results found in the literature for these databases. Idioma: Inglés DOI: 10.1186/s13636-020-00172-6 Año: 2020 Publicado en: EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING 2020 (2020), 5 [19 pp.] ISSN: 1687-4714 Factor impacto JCR: 1.558 (2020) Categ. JCR: ENGINEERING, ELECTRICAL & ELECTRONIC rank: 198 / 273 = 0.725 (2020) - Q3 - T3 Categ. JCR: ACOUSTICS rank: 20 / 32 = 0.625 (2020) - Q3 - T2 Factor impacto SCIMAGO: 0.259 - Electrical and Electronic Engineering (Q3) - Acoustics and Ultrasonics (Q3)