RISS 학술연구정보서비스

검색
다국어 입력

http://chineseinput.net/에서 pinyin(병음)방식으로 중국어를 변환할 수 있습니다.

변환된 중국어를 복사하여 사용하시면 됩니다.

예시)
  • 中文 을 입력하시려면 zhongwen을 입력하시고 space를누르시면됩니다.
  • 北京 을 입력하시려면 beijing을 입력하시고 space를 누르시면 됩니다.
닫기
    인기검색어 순위 펼치기

    RISS 인기검색어

      검색결과 좁혀 보기

      선택해제
      • 좁혀본 항목 보기순서

        • 원문유무
        • 음성지원유무
        • 원문제공처
          펼치기
        • 등재정보
          펼치기
        • 학술지명
          펼치기
        • 주제분류
          펼치기
        • 발행연도
          펼치기
        • 작성언어

      오늘 본 자료

      • 오늘 본 자료가 없습니다.
      더보기
      • 무료
      • 기관 내 무료
      • 유료
      • KCI등재

        베이시안 기법과 선택적 음성특징 추출을 융합한 음성 인식 성능 향상

        황재천 한국융합학회 2016 한국융합학회논문지 Vol.7 No.6

        일반적인 어휘 인식 시스템은 백색 잡음과 음성을 인식하는 환경에서 여러 음성의 혼재되어 정확한 음성 을 인식하지 못하고 있다. 따라서 본 논문은 효율적인 음성 인식을 위해 잡음 음성으로 부터 원하는 음성만 선택적 으로 추출하기 위한 방법과 베이시안 기법을 융합 방법을 제안한다. 음성의 선택적 추출을 위해 필터 뱅크 주파수 응답 계수를 사용한다. 하며, 이를 위해 모든 가능한 두 관측치의 조합에 대해 변수 관측치를 사용하며, 음성 신호 정보를 가지고 선택적 음성 특징 추출을 위해 잡음은 출력에 대한 에너지 비율을 구한다. 이것은 음성 특징을 추출 하는 방법을 제안하며, 이를 베이시안 기법의 어휘 인식을 융합하여 잡음을 제거하고 인식률을 향상시켰다. 본 논문에서 기존의 HMM과 CHMM 방법과 비교한 결과 잡음 환경에서의 인식률이 2.3% 향상됨을 확인하였다 Voice recognition systems which use a white noise and voice recognition environment are not correct voice recognition with variable voice mixture. Therefore in this paper, we propose a method using the convergence of Bayesian technique and selecting voice for effective voice recognition. we make use of bank frequency response coefficient for selective voice extraction, Using variables observed for the combination of all the possible two observations for this purpose, and has an voice signal noise information to the speech characteristic extraction selectively is obtained by the energy ratio on the output. It provide a noise elimination and recognition rates are improved with combine voice recognition of bayesian methode. The result which we confirmed that the recognition rate of 2.3% is higher than HMM and CHMM methods in vocabulary recognition, respectively.

      • FPGA Design of Voice Enabled Ignition using G279 for Modal Based Speech Compression

        Kelan McLean,Marcus Lloyde George 보안공학연구지원센터 2016 International Journal of Grid and Distributed Comp Vol.9 No.6

        Voice enabled ignition combines the speaker recognition and word recognition aspects of speech recognition. It replaces the function of a key in the starting of the ignition system of a car. An Fpga design incorporates the required components of a generic speech recognition system and uses the unique capabilities of hardware in term of parallelism to improve performance. The compression of speech for storage and playback was facilitated by the usage of the G729 standard for compression of speech.

      • KCI등재

        문서 편집 접근성 향상을 위한 음성 명령기반 모바일 어플리케이션 개발

        박주현,박세아,이무늬,임순범 한국멀티미디어학회 2018 멀티미디어학회논문지 Vol.21 No.11

        Voice Command systems are important means of ensuring accessibility to digital devices for use in situations where both hands are not free or for people with disabilities. Interests in services using speech recognition technology have been increasing. In this study, we developed a mobile writing application using voice recognition and voice command technology which helps people create and edit documents easily. This application is characterized by the minimization of the touch on the screen and the writing of memo by voice. We have systematically designed a mode to distinguish voice writing and voice command so that the writing and execution system can be used simultaneously in one voice interface. It provides a shortcut function that can control the cursor by voice, which makes document editing as convenient as possible. This allows people to conveniently access writing applications by voice under both physical and environmental constraints.

      • Multimodal Biometric Recognition System for Cloud Robots

        Shuqing Tian,Sung Gyu Im,Suk Gyu Lee 보안공학연구지원센터 2015 International Journal of Security and Its Applicat Vol.9 No.7

        This paper presents a Multimodal Biometric Recognition System (MBRS) which is capable of integrating various biometric information for person recognition. The MBRS is deployed as a cloud server and provides person recognition service for smart robots. Through the experiments based on multimodal biometric traits, the fact that the multimodal biometric recognition performs better than individual biometric recognition has been proved. In our approach, the implementation of a multimodal biometric recognition system based on face recognition system and voice recognition system is proposed. The MBRS provides the possibility of integrating multi biometric subsystems to do recognition. Even more, since the MBRS is deployed as a cloud server, the public interfaces were provided for the robots to do real-time person recognition. The experimental results show that the MBRS outperforms any individual face recognition subsystem and voice recognition subsystem.

      • KCI등재

        화자식별 기반의 AI 음성인식 서비스에 대한 사이버 위협 분석

        홍천호 ( Chunho Hong ),조영호 ( Youngho Cho ) 한국인터넷정보학회 2021 인터넷정보학회논문지 Vol.22 No.6

        음성인식(ASR: Automatic Speech Recognition)은 사람의 말소리를 음성 신호로 분석하고, 문자열로 자동 변화하여 이해하는 기술이다. 초기 음성인식 기술은 하나의 단어를 인식하는 것을 시작으로 두 개 이상의 단어로 구성된 문장을 인식하는 수준까지 진화하였다. 실시간 음성 대화에 있어 높은 인식률은 자연스러운 정보전달의 편리성을 극대화하여 그 적용 범위를 확장하고 있다. 반면에, 음성인식 기술의 활발한 적용에 따라 관련된 사이버 공격과 위협에 대한 우려 역시 증가하고 있다. 기존 연구를 살펴보면, 자동화자 식별(ASV: Automatic Speaker Verification) 기법의 고안과 정확성 향상 등 기술 발전 자체에 관한 연구는 활발히 이루어지고 있으나, 실생활에 적용되고 있는 음성인식 서비스의 자동화자 식별 기술에 대한 사이버 공격 및 위협에 관한 분석연구는 다양하고 깊이 있게 수행되지 않고 있다. 본 연구에서는 자동화자 식별 기술을 갖춘 AI 음성인식 서비스를 대상으로 음성 주파수와 음성속도를 조작하여 음성인증을 우회하는 사이버 공격 모델을 제안하고, 상용 스마트폰의 자동화자 식별 체계를 대상으로 실제 실험을 통해 사이버 위협을 분석한다. 이를 통해 관련 사이버 위협의 심각성을 알리고 효과적인 대응 방안에 관한 연구 관심을 높이고자 한다. Automatic Speech Recognition(ASR) is a technology that analyzes human speech sound into speech signals and then automatically converts them into character strings that can be understandable by human. Speech recognition technology has evolved from the basic level of recognizing a single word to the advanced level of recognizing sentences consisting of multiple words. In real-time voice conversation, the high recognition rate improves the convenience of natural information delivery and expands the scope of voice-based applications. On the other hand, with the active application of speech recognition technology, concerns about related cyber attacks and threats are also increasing. According to the existing studies, researches on the technology development itself, such as the design of the Automatic Speaker Verification(ASV) technique and improvement of accuracy, are being actively conducted. However, there are not many analysis studies of attacks and threats in depth and variety. In this study, we propose a cyber attack model that bypasses voice authentication by simply manipulating voice frequency and voice speed for AI voice recognition service equipped with automated identification technology and analyze cyber threats by conducting extensive experiments on the automated identification system of commercial smartphones. Through this, we intend to inform the seriousness of the related cyber threats and raise interests in research on effective countermeasures.

      • KCI등재

        음성인식기반 스마트 의료조명 제어시스템

        김민규,이수인,조현길 대한임베디드공학회 2013 대한임베디드공학회논문지 Vol.8 No.3

        A voice recognition technology as a technology fundament plays an important role in medical devices with smart functions. This paper describes the implementation of a control system that can be utilized as a part of illumination equipment for medical applications (IEMA) based on a voice recognition. The control system can essentially be divided into five parts, the microphone, training part, recognition part, memory part, and control part. The system was implemented using the RSC-4x evaluation board which is included the micro-controller for voice recognition. To investigate the usefulness of the implemented control system, the experiments of the recognition rate was carried out according to the input distance for voice recognition. As a result, the recognition rate of the control system was more than 95% within a distance between 0.5 and 2m. The result verified that the implemented control system performs well as the smart control system based for an IEMA.

      • KCI등재

        Voice Recognition Technologies: Comparative Analysis and Potential Challenges in Future Implementation

        Mamta Devi,Kabir Md Shahriar,고일상 한국인터넷전자상거래학회 2023 인터넷전자상거래연구 Vol.23 No.6

        This comprehensive study delves into the current landscape of Voice Recognition Technologies (VRT), focusing on trends, implementation practices, and the challenges faced by users. Google Assistant, Amazon’s Alexa, and Apple’s Siri serve as pivotal subjects, providing a foundational basis for examination. The research commences with an extensive literature review, meticulously filtering and streamlining relevant literature on the subject. Employing a qualitative approach, the study conducts open-ended interviews, posing ten main questions directly linked to the central theme. The qualitative analysis is complemented by academic evidence in the final chapters, ensuring a robust foundation for the study's outcomes. The findings underscore the significant technological strides made by voice recognition; however, challenges persist, including hands-free connectivity, privacy and security concerns, and the issue of voice imprinting. These challenges, identified through user experiences and theoretical insights, illuminate the path for future research and development, guiding the evolution of VRT towards a more seamless, secure, and user-friendly future. This research not only contributes to academic knowledge but also holds practical implications for technology experts, policymakers, and businesses, paving the way for informed decision-making and innovation in the dynamic field of voice recognition.

      • KCI등재

        Development of a Work Management System Based on Speech and Speaker Recognition

        게이뷸라예프 압둘라지즈,유누소프 자홍길,김태형 대한임베디드공학회 2021 대한임베디드공학회논문지 Vol.16 No.3

        Voice interface can not only make daily life more convenient through artificial intelligence speakers but also improve the working environment of the factory. This paper presents a voice-assisted work management system that supports both speech and speaker recognition. This system is able to provide machine control and authorized worker authentication by voice at the same time. We applied two speech recognition methods, Google’s Speech application programming interface (API) service, and DeepSpeech speech-to-text engine. For worker identification, the SincNet architecture for speaker recognition was adopted. We implemented a prototype of the work management system that provides voice control with 26 commands and identifies 100 workers by voice. Worker identification using our model was almost perfect, and the command recognition accuracy was 97.0% in Google API after post- processing and 92.0% in our DeepSpeech model.

      • KCI등재

        Development of a Work Management System Based on Speech and Speaker Recognition

        Gaybulayev, Abdulaziz,Yunusov, Jahongir,Kim, Tae-Hyong Institute of Embedded Engineering of Korea 2021 대한임베디드공학회논문지 Vol.16 No.3

        Voice interface can not only make daily life more convenient through artificial intelligence speakers but also improve the working environment of the factory. This paper presents a voice-assisted work management system that supports both speech and speaker recognition. This system is able to provide machine control and authorized worker authentication by voice at the same time. We applied two speech recognition methods, Google's Speech application programming interface (API) service, and DeepSpeech speech-to-text engine. For worker identification, the SincNet architecture for speaker recognition was adopted. We implemented a prototype of the work management system that provides voice control with 26 commands and identifies 100 workers by voice. Worker identification using our model was almost perfect, and the command recognition accuracy was 97.0% in Google API after post- processing and 92.0% in our DeepSpeech model.

      • KCI등재

        고령화 사회를 위한 음성 인식 챗봇 시스템 : 기술 개발과 맞춤형 UI/UX 설계

        정윤지,유민성,오주영,황현석,허원회,Yun-Ji Jeong,Min-Seong Yu,Joo-Young Oh,Hyeon-Seok Hwang,Won-Whoi Hun 한국인터넷방송통신학회 2024 한국인터넷방송통신학회 논문지 Vol.24 No.4

        본 연구는 고령화 사회의 노년층 우울증과 고독감 문제를 해결하기 위해 음성 인식 챗봇 시스템을 개발하였다. 이 시스템은 Whisper 모델, GPT 2.5, XTTS2를 활용하여 고성능 음성 인식과 자연어 처리, 텍스트-음성 변환 기능을 제공한다. 사용자는 이를 통해 감정과 상태를 표현하고 적절한 반응을 얻을 수 있으며, 지인의 목소리를 이용한 음성인식 기능으로 친숙함과 안정감을 느낄 수 있다. UX/UI는 스마트 시니어 세대의 인지 반응과 시력 저하, 운동 능력 제약 등을 고려하여 설계되었다. 명도와 선명도가 높은 색상, 가독성이 좋은 서체등을 활용하여 고령자의 사용 편의성을 높였다.이 연구는 음성 기반 인터페이스를 통해 노년층의 삶의 질 향상에 기여할 것으로 기대된다. This study developed a voice recognition chatbot system to address depression and loneliness among the elderly in an aging society. The system utilizes the Whisper model, GPT 2.5, and XTTS2 to provide high-performance voice recognition, natural language processing, and text-to-speech conversion. Users can express their emotions and states and receive appropriate responses, with voice recognition functionality using familiar voices for comfort and reassurance. The UX/UI design considers the cognitive responses, visual impairments, and physical limitations of the smart senior generation, using high contrast colors and readable fonts for enhanced usability. This research is expected to improve the quality of life for the elderly through voice-based interfaces.

      연관 검색어 추천

      이 검색어로 많이 본 자료

      활용도 높은 자료

      해외이동버튼