1:"$Sreact.fragment" 2:I[97367,["/_next/static/chunks/ff1a16fafef87110.js","/_next/static/chunks/247eb132b7f7b574.js"],"OutletBoundary"] 3:"$Sreact.suspense" :HL["/portfolio/speech-emotion.png","image"] 0:{"buildId":"HmrLMW_hY8v7b2jIEIUBW","rsc":["$","$1","c",{"children":[["$","main",null,{"className":"container-default py-12","children":[["$","h1",null,{"className":"text-4xl font-bold mb-6 text-white","children":"Speech Emotion Recognition"}],["$","img",null,{"src":"/portfolio/speech-emotion.png","alt":"Speech Emotion Recognition","className":"rounded-xl mb-6 w-full h-[32rem] object-cover border border-white/10"}],["$","section",null,{"className":"mb-8","children":[["$","p",null,{"className":"text-lg text-white/80 mb-4","children":"An advanced NLP and Audio analysis AI utilizing Gated Recurrent Units (GRU). It extracts complex human emotional insights directly from speech nuances, providing deep behavioral analytics."}],["$","ul",null,{"className":"list-disc ml-6 text-lg text-white/70 mb-4","children":[["$","li",null,{"children":"Raw audio waveform processing and feature extraction"}],["$","li",null,{"children":"GRU-based temporal sequence modeling for shifting emotions"}],["$","li",null,{"children":"Multilingual tone and sentiment detection capabilities"}],["$","li",null,{"children":"Scalable architecture for processing bulk customer service calls"}]]}],["$","div",null,{"className":"mb-8","children":[["$","h2",null,{"className":"text-2xl font-semibold text-white mb-2","children":"Project Impact"}],["$","p",null,{"className":"text-white/70 mb-2","children":"Revolutionized customer service quality assurance by automatically flagging calls with high frustration or anger, allowing immediate intervention and de-escalation by human managers."}]]}]]}]]}],null,["$","$L2",null,{"children":["$","$3",null,{"name":"Next.MetadataOutlet","children":"$@4"}]}]]}],"loading":null,"isPartial":false} 4:null