{"id":37705,"date":"2025-01-13T16:36:49","date_gmt":"2025-01-13T06:36:49","guid":{"rendered":"https:\/\/myassignmenthelp.info\/assignments\/?p=37705"},"modified":"2025-01-13T16:36:50","modified_gmt":"2025-01-13T06:36:50","slug":"emerging-technologies-workshop-6-2352685","status":"publish","type":"post","link":"https:\/\/myassignmenthelp.info\/assignments\/emerging-technologies-workshop-6-2352685\/","title":{"rendered":"Emerging Technologies Workshop 6-2352685"},"content":{"rendered":"<p><div class=\"ppw-restricted-content\"><\/p>\n\n\n\n<p>Task 3 (Text Generation):<br>a) What is tokenization?<br>Basically, tokenization deals with the segmenting process of a whole text into minimal units such<br>as words, subwords, and\/or even characters for a machine to work on (Islam et al., 2021).<br>b) Why do we need to tokenize the input text?<br>Text tokenization is essential because neural networks can&#8217;t process raw text directly. It converts<br>text into numerical sequences that machines can understand and process (Islam et al., 2021).<br>c) How can we tokenize the input text?<br>Word-based tokenization: Splitting by spaces or punctuation.<br>Subword tokenization: Methods like BPE (Byte Pair Encoding) or WordPiece.<br>Character-based tokenization: Splitting text into individual characters.<br>Sentence tokenization: Dividing text into sentences (Islam et al., 2021).<br>Task 4: NLP and Sequence Models<br>a) What is Natural Language Processing (NLP)?<br>Natural Language Processing alludes to a technique of using computational methods to make<br>computers generate and comprehend human languages (Min et al., 2023).<br>b) Sequence Models in Deep Learning:<br>Sequence models process data where order matters, like text or time series. They maintain<br>context across inputs, essential for understanding language structure and meaning.<br>c) Sequence Model Types:<br>One-to-One: Single input to single output (e.g., standard classification tasks).<br>One-to-Many: Single input generating multiple outputs (e.g., image captioning).<br>3<br>Many-to-One: Multiple inputs producing a single output (e.g., sentiment analysis)..<br>Many-to-Many: Multiple inputs generating multiple outputs (e.g., language translation) (Smith<br>et al., 2022).<br>Task 6: Image Generation Challenges<br>a) Diffusion Model Challenges:<br>High Computational Cost: Training and inference require significant computational resources<br>due to the iterative denoising process.<br>Slow Generation Speed: Diffusion models are slower compared to GANs because of their<br>sequential denoising nature (Huang et al., 2024).<br>b) Conditional GAN Operation:<br>Conditional GANs incorporate additional input (e.g., labels or conditions) alongside random<br>noise to control generation. The discriminator evaluates both the generated image and the<br>conditioning information, enabling targeted content generation.<br>Task 7: Advanced GenAI Concepts<br>a) Parallelization:<br>Parallelization is the process of splitting computations across multiple processors or GPUs to<br>perform tasks simultaneously, increasing efficiency and speed (Luitse and Denkena 2021).<br>b) End-to-End Deep Learning:<br>The system processes raw input directly to the desired output without manual feature engineering<br>and learns all transformations automatically (Qi et al., 2023).<br>c) Video Generation Challenges:<br>Temporal Consistency: Ensuring realistic motion and seamless transitions across frames while<br>maintaining coherence.<br>4<br>High Computational Requirements: Generating multiple frames while preserving spatialtemporal relationships demands significant resources.<br>5<br>References<br>Huang, Y., Huang, J., Liu, Y., Yan, M., Lv, J., Liu, J., Xiong, W., Zhang, H., Chen, S. and Cao,<br>L. (2024). Diffusion model-based image editing: A survey. arXiv preprint arXiv:2402.17525.<br>Islam, T., Hossain, M. and Arefin, M.F. (2021, December). Comparative analysis of different<br>text summarization techniques using enhanced tokenization. In 2021 3rd International<br>Conference on Sustainable Technologies for Industry 4.0 (STI) (pp. 1-6). IEEE.<br>Luitse, D. and Denkena, W. (2021). The great transformer: Examining the role of large language<br>models in the political economy of AI. Big Data &amp; Society, 8(2), p.20539517211047734.<br>Min, B., Ross, H., Sulem, E., Veyseh, A.P.B., Nguyen, T.H., Sainz, O., Agirre, E., Heintz, I. and<br>Roth, D. (2023). Recent advances in natural language processing via large pre-trained language<br>models: A survey. ACM Computing Surveys, 56(2), pp.1-40.<br>Smith, J.T., Warrington, A. and Linderman, S.W. (2022). Simplified state space layers for<br>sequence modeling. arXiv preprint arXiv:2208.04933.<br>Qi, M., Shi, Y., Qi, Y., Ma, C., Yuan, R., Wu, D. and Shen, Z.J. (2023). A practical end-to-end<br>inventory management model with deep learning. Management Science, 69(2), pp.759-773.<\/p>\n\n\n\n<p><\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Use IoT devices to monitor player movements in sports such as:<br>o Soccer: Determination of player positioning and their levels of energy.<br>o Basketball: Collect data on the heights of jumps and player ability to move<br>quickly during the game.<br>o Rugby: Measure the efficiency of sprints and tackling methods.<br>Impact:<\/li>\n\n\n\n<li>Enhances team results by establishing player strengths and weaknesses.<\/li>\n\n\n\n<li>Offering coaches with comprehensive game analytics (Yuan et al., 2022).<br>Future Research:<\/li>\n\n\n\n<li>Integration with drones for synchronized aerial and ground-level motion analysis.<\/li>\n\n\n\n<li>Player performance assessments using artificial intelligence for creation of individual<br>training programmes.<br>Example 2: Fall Prevention for Elderly Care<br>Application:<\/li>\n\n\n\n<li>IoT wearables detect the gait patterns of elderly people and monitor them.<\/li>\n\n\n\n<li>Signal caregivers in cases of detection of irregular movements or when a fall may occur.<br>Impact:<\/li>\n\n\n\n<li>Enhances safety and independence for the elderly.<\/li>\n\n\n\n<li>Reduces the cost of health care by preventing fall injuries (Kulurkar et al., 2023).<br>Future Research:<\/li>\n\n\n\n<li>Advanced AI algorithms to predict fall risks based on historical motion data.<\/li>\n\n\n\n<li>Smart home system integrations for automatic responses in case of emergencies.<br>3<br>Example 3: Motion Monitoring in Fitness Apps<br>Application:<\/li>\n\n\n\n<li>Wearable fitness devices can track workout efficiency in various activities including the<br>following:<br>o Yoga: Evaluate the correctness and stability of the pose.<br>o Weightlifting: Keep track of the posture and range of motion.<br>o Cycling: Cadence and energy efficiency analysis.<br>Impact:<\/li>\n\n\n\n<li>Personalizes fitness programs with individual performance data.<\/li>\n\n\n\n<li>Enhances engagement through feedback provided in real time and indicating progress.<br>Future Research:<\/li>\n\n\n\n<li>Integration of virtual reality into fitness for better immersion.<\/li>\n\n\n\n<li>New, wearable designs that offer unhindered motion tracking for several activities (Cai,<br>2022).<br>Task 2 \u2013 Individual Research: Tools for Motion Analysis<br>Summary of Findings<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Other Tools for Motion Analysis in Sports and Health Contexts<br>a. IMU (Inertial Measurement Unit) Sensors<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>What It Is: Small devices for measuring acceleration, angular speed, and orientation.<\/li>\n\n\n\n<li>Applications:<br>o Sports: Monitor biomechanics when it comes to exercising, for instance, jogging,<br>swimming or cycling.<br>o Healthcare: Controlling the mobility of the concerned patient during the<br>rehabilitation exercise (Samatas and Pachidis, 2022).<\/li>\n\n\n\n<li>Example Tools:<br>4<br>o Perception Neuron: An inexpensive wearable system for full body tracking.<br>o Xsens: High-precision sensors for biomechanical investigations.<br>b. Marker-Based Motion Capture Systems<\/li>\n\n\n\n<li>What It Is: Systems that rely on camera to detect reflectance markers of the body<br>(Kanko et al., 2021).<\/li>\n\n\n\n<li>Applications:<br>o Sports: Assess some intricate patterns in either gymnastics or dance.<br>o Healthcare: Applied in the clinical evaluation of walking pattern and in the<br>design of artificial limbs.<\/li>\n\n\n\n<li>Example Tools:<br>o Vicon: The high-accuracy systems used in biomechanics are motion capture<br>systems.<br>o Qualisys: The latest technological equipment for three-dimensional motion<br>tracking in special laboratories.<br>c. Markerless Motion Capture<\/li>\n\n\n\n<li>What It Is: AI-based systems that doesn&#8217;t take use of physical characteristics (Kanko et<br>al., 2021).<\/li>\n\n\n\n<li>Applications:<br>o Sports: Evaluate current athletic performance during a certain sport.<br>o Healthcare: Management of natural movements during rehabilitation of patient<br>that had a stroke.<\/li>\n\n\n\n<li>Example Tools:<br>o OpenPose: Open-source tool that estimates pose.<br>o Kinovea: Free software analyzing motion from videos<br>5<br>d. Pressure and Force Platforms<\/li>\n\n\n\n<li>What It Is: Tools that measure ground reaction forces during movements.<\/li>\n\n\n\n<li>Applications:<br>o Sports: Evaluation of jumps and sprint starts.<br>o Healthcare: Assess balance and fall risks in elderly individuals (Wang et al.,<br>2022).<\/li>\n\n\n\n<li>Example Tools:<br>o AMTI: Force plates for biomechanical research.<br>o Bertec: Specific systems intended for gait and balance analysis.<br>e. Smart Wearables<\/li>\n\n\n\n<li>What It Is: The wearable devices popular among consumers as fitness and health<br>tracking gadgets.<\/li>\n\n\n\n<li>Applications:<br>o Sports: Monitor metrics such as, heart rate, steps, and general body motions.<br>o Healthcare: Keep track of physical activity as well as quality of sleep.<\/li>\n\n\n\n<li>Example Tools:<br>o Garmin and Fitbit: Popular fitness trackers.<br>o WHOOP: A performance and recovery monitoring wearable.<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Research Trends and Directions<br>a. Real-Time Feedback Systems<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Tools are increasingly focused on delivering instant feedback to improve performance or<br>recovery.<\/li>\n\n\n\n<li>Examples: Smart glasses for information purposes, and vibrating feedback devices in<br>wearable technology.<br>6<br>b. AI Integration<\/li>\n\n\n\n<li>AI improves motion data interpretation by enabling predictions and further analysis.<\/li>\n\n\n\n<li>Examples: Machine learning models for estimating the probability of an injury from the<br>movement profile.<br>c. Remote Monitoring<\/li>\n\n\n\n<li>Due to IoT, patient or athlete tracking systems are possible through cloud platforms.<br>d. Immersive Visualization<\/li>\n\n\n\n<li>Utilization of augmented and virtual reality for the visualization of biomechanical data.<br>7<br>References<br>Cai, H. (2022). Application of intelligent real-time image processing in fitness motion detection<br>under internet of things. The Journal of Supercomputing, 78(6), pp.7788-7804.<br>Kanko, R.M., Laende, E.K., Davis, E.M., Selbie, W.S. and Deluzio, K.J. (2021). Concurrent<br>assessment of gait kinematics using marker-based and markerless motion capture. Journal of<br>biomechanics, 127, p.110665.<br>Kulurkar, P., kumar Dixit, C., Bharathi, V.C., Monikavishnuvarthini, A., Dhakne, A. and Preethi,<br>P. (2023). AI based elderly fall prediction system using wearable sensors: A smart home-care<br>technology with IOT. Measurement: Sensors, 25, p.100614.<br>Samatas, G.G. and Pachidis, T.P. (2022). Inertial measurement units (imus) in mobile robots<br>over the last five years: A review. Designs, 6(1), p.17.<br>Yuan, Y., Lu, Z., Yang, Z., Jian, M., Wu, L., Li, Z. and Liu, X. (2022). Key frame extraction<br>based on global motion statistics for team-sport videos. Multimedia Systems, 28(2), pp.387-401.<br>Wang, J., Zhu, Y., Wu, Z., Zhang, Y., Lin, J., Chen, T., Liu, H., Wang, F. and Sun, L. (2022).<br>Wearable multichannel pulse condition monitoring system based on flexible pressure sensor<br>arrays. Microsystems &amp; Nanoengineering, 8(1), p.16.<\/li>\n<\/ul>\n\n\n\n<p>1<br>Week 5 Exercises<br>(By Name)<br>Course<br>University<br>Professor<br>State<br>City<br>Date<br>2<br>Exercises and Answers<br>Exercise 1: Application Area of Computer Vision<br>Favorite Application: Tesla&#8217;s Autonomous Driving<br>Tesla employs computer vision in its Autopilot and Full Self-Driving (FSD) systems to enable<br>real-time navigation with minimal human intervention. Key components include neural networks<br>trained on camera feeds, ultrasonic sensors, and radar.<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Examples:<br>o Tesla\u2019s Smart Summon feature autonomously drives cars to owners in parking<br>lots.<br>o AI Day presentations demonstrate the vision-first approach, avoiding LiDAR and<br>focusing on camera-based perception.<br>o Dojo Supercomputer trains its neural networks using millions of driving videos.<\/li>\n\n\n\n<li>Videos and research on Tesla\u2019s capabilities highlight advancements in lane detection,<br>object prediction, and traffic analysis.<br>Exercise 2: Deep Fake Videos<br>Observations:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Best Video: The Barack Obama deep fake video created by Jordan Peele. It convincingly<br>showcased how AI can manipulate speech and facial expressions.<\/li>\n\n\n\n<li>How They Did It: This video was created using GANs (Generative Adversarial<br>Networks) and facial mapping technology. Specific tools like DeepFaceLab or customtrained models were used to replicate Obama\u2019s speech patterns and expressions.<\/li>\n\n\n\n<li>Time and Hardware: Likely required weeks of training on high-end GPUs, given its<br>quality and precision.<br>3<\/li>\n\n\n\n<li>Realism: While highly realistic, close inspection reveals slight lip-sync inconsistencies<br>and unnatural transitions in facial movements.<br>Exercise 3: Wacky Object Variations<br>Category: Potted Plants<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Found images include:<br>o A plant shaped like a cat.<br>o A pot designed as a robot holding a small cactus.<br>o A plant growing from a shoe instead of a traditional pot.<br>Exercise 4: Rise of Deep Learning<br>Key Points:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Trends in GPU Price Performance:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>GPUs have improved dramatically in computational power while becoming more<br>affordable, enabling broader adoption of deep learning.<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>NVIDIA and AI Dominance:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>NVIDIA\u2019s GPUs revolutionized AI training, surpassing CPU-based methods,<br>with partnerships accelerating AI integration in diverse industries.<br>Exercise 5: Object Detection<\/li>\n\n\n\n<li>Tested images include the unusual plants from Exercise 3.<\/li>\n\n\n\n<li>Observations:<br>o Accurate detection of plant-like shapes.<br>o Unique pots were classified as unrelated objects like furniture.<br>4<br>Exercise 6: Image Classification<br>Observations:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Single Object: A clean headshot was identified with high confidence and minimal<br>ambiguity.<\/li>\n\n\n\n<li>Cluttered Scene: A coffee cup amidst a desk setup confused the model, leading to<br>multiple low-confidence classifications.<br>Exercise 7: One-Stage vs Two-Stage Models<br>Models:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>YOLO (One-Stage):<br>o Speed: ~45 FPS<br>o Accuracy: ~55 mAP on COCO dataset.<\/li>\n\n\n\n<li>Faster-RCNN (Two-Stage):<br>o Speed: ~5 FPS<br>o Accuracy: ~60-65 mAP on COCO dataset.<br>Comparison: YOLO prioritizes speed for real-time applications, while Faster-RCNN offers<br>higher accuracy at the cost of slower inference.<br>Bonus Insights:<\/li>\n\n\n\n<li>DETR achieves mAP scores around 65-70 on COCO but has slower inference times (~10<br>FPS).<\/li>\n<\/ul>\n\n\n\n<p>Exercises and Answers<br>Exercise 1: Application Area of Computer Vision<br>Favorite Application: Tesla&#8217;s Autonomous Driving<br>Tesla employs computer vision in its Autopilot and Full Self-Driving (FSD) systems to enable<br>real-time navigation with minimal human intervention. Key components include neural networks<br>trained on camera feeds, ultrasonic sensors, and radar.<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Examples:<br>o Tesla\u2019s Smart Summon feature autonomously drives cars to owners in parking<br>lots.<br>o AI Day presentations demonstrate the vision-first approach, avoiding LiDAR and<br>focusing on camera-based perception.<br>o Dojo Supercomputer trains its neural networks using millions of driving videos.<\/li>\n\n\n\n<li>Videos and research on Tesla\u2019s capabilities highlight advancements in lane detection,<br>object prediction, and traffic analysis.<br>Exercise 2: Deep Fake Videos<br>Observations:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Best Video: The Barack Obama deep fake video created by Jordan Peele. It convincingly<br>showcased how AI can manipulate speech and facial expressions.<\/li>\n\n\n\n<li>How They Did It: This video was created using GANs (Generative Adversarial<br>Networks) and facial mapping technology. Specific tools like DeepFaceLab or customtrained models were used to replicate Obama\u2019s speech patterns and expressions.<\/li>\n\n\n\n<li>Time and Hardware: Likely required weeks of training on high-end GPUs, given its<br>quality and precision.<br>3<\/li>\n\n\n\n<li>Realism: While highly realistic, close inspection reveals slight lip-sync inconsistencies<br>and unnatural transitions in facial movements.<br>Exercise 3: Wacky Object Variations<br>Category: Potted Plants<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Found images include:<br>o A plant shaped like a cat.<br>o A pot designed as a robot holding a small cactus.<br>o A plant growing from a shoe instead of a traditional pot.<br>Exercise 4: Rise of Deep Learning<br>Key Points:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Trends in GPU Price Performance:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>GPUs have improved dramatically in computational power while becoming more<br>affordable, enabling broader adoption of deep learning.<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>NVIDIA and AI Dominance:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>NVIDIA\u2019s GPUs revolutionized AI training, surpassing CPU-based methods,<br>with partnerships accelerating AI integration in diverse industries.<br>Exercise 5: Object Detection<\/li>\n\n\n\n<li>Tested images include the unusual plants from Exercise 3.<\/li>\n\n\n\n<li>Observations:<br>o Accurate detection of plant-like shapes.<br>o Unique pots were classified as unrelated objects like furniture.<br>4<br>Exercise 6: Image Classification<br>Observations:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Single Object: A clean headshot was identified with high confidence and minimal<br>ambiguity.<\/li>\n\n\n\n<li>Cluttered Scene: A coffee cup amidst a desk setup confused the model, leading to<br>multiple low-confidence classifications.<br>Exercise 7: One-Stage vs Two-Stage Models<br>Models:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>YOLO (One-Stage):<br>o Speed: ~45 FPS<br>o Accuracy: ~55 mAP on COCO dataset.<\/li>\n\n\n\n<li>Faster-RCNN (Two-Stage):<br>o Speed: ~5 FPS<br>o Accuracy: ~60-65 mAP on COCO dataset.<br>Comparison: YOLO prioritizes speed for real-time applications, while Faster-RCNN offers<br>higher accuracy at the cost of slower inference.<br>Bonus Insights:<\/li>\n\n\n\n<li>DETR achieves mAP scores around 65-70 on COCO but has slower inference times (~10<br>FPS).<\/li>\n<\/ul>\n\n\n\n<p>Exercise 1: Application Area of Computer Vision<br>Favorite Application: Tesla&#8217;s Autonomous Driving<br>Tesla employs computer vision in its Autopilot and Full Self-Driving (FSD) systems to enable<br>real-time navigation with minimal human intervention. Key components include neural networks<br>trained on camera feeds, ultrasonic sensors, and radar.<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Examples:<br>o Tesla\u2019s Smart Summon feature autonomously drives cars to owners in parking<br>lots.<br>o AI Day presentations demonstrate the vision-first approach, avoiding LiDAR and<br>focusing on camera-based perception.<br>o Dojo Supercomputer trains its neural networks using millions of driving videos.<\/li>\n\n\n\n<li>Videos and research on Tesla\u2019s capabilities highlight advancements in lane detection,<br>object prediction, and traffic analysis.<br>Exercise 2: Deep Fake Videos<br>Observations:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Best Video: The Barack Obama deep fake video created by Jordan Peele. It convincingly<br>showcased how AI can manipulate speech and facial expressions.<\/li>\n\n\n\n<li>How They Did It: This video was created using GANs (Generative Adversarial<br>Networks) and facial mapping technology. Specific tools like DeepFaceLab or customtrained models were used to replicate Obama\u2019s speech patterns and expressions.<\/li>\n\n\n\n<li>Time and Hardware: Likely required weeks of training on high-end GPUs, given its<br>quality and precision.<br>3<\/li>\n\n\n\n<li>Realism: While highly realistic, close inspection reveals slight lip-sync inconsistencies<br>and unnatural transitions in facial movements.<br>Exercise 3: Wacky Object Variations<br>Category: Potted Plants<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Found images include:<br>o A plant shaped like a cat.<br>o A pot designed as a robot holding a small cactus.<br>o A plant growing from a shoe instead of a traditional pot.<br>Exercise 4: Rise of Deep Learning<br>Key Points:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Trends in GPU Price Performance:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>GPUs have improved dramatically in computational power while becoming more<br>affordable, enabling broader adoption of deep learning.<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>NVIDIA and AI Dominance:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>NVIDIA\u2019s GPUs revolutionized AI training, surpassing CPU-based methods,<br>with partnerships accelerating AI integration in diverse industries.<br>Exercise 5: Object Detection<\/li>\n\n\n\n<li>Tested images include the unusual plants from Exercise 3.<\/li>\n\n\n\n<li>Observations:<br>o Accurate detection of plant-like shapes.<br>o Unique pots were classified as unrelated objects like furniture.<br>4<br>Exercise 6: Image Classification<br>Observations:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Single Object: A clean headshot was identified with high confidence and minimal<br>ambiguity.<\/li>\n\n\n\n<li>Cluttered Scene: A coffee cup amidst a desk setup confused the model, leading to<br>multiple low-confidence classifications.<br>Exercise 7: One-Stage vs Two-Stage Models<br>Models:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>YOLO (One-Stage):<br>o Speed: ~45 FPS<br>o Accuracy: ~55 mAP on COCO dataset.<\/li>\n\n\n\n<li>Faster-RCNN (Two-Stage):<br>o Speed: ~5 FPS<br>o Accuracy: ~60-65 mAP on COCO dataset.<br>Comparison: YOLO prioritizes speed for real-time applications, while Faster-RCNN offers<br>higher accuracy at the cost of slower inference.<br>Bonus Insights:<\/li>\n\n\n\n<li>DETR achieves mAP scores around 65-70 on COCO but has slower inference times (~10<br>FPS).<\/li>\n<\/ul>\n\n\n\n<p>Exercises and Answers<br>Exercise 1: Application Area of Computer Vision<br>Favorite Application: Tesla&#8217;s Autonomous Driving<br>Tesla employs computer vision in its Autopilot and Full Self-Driving (FSD) systems to enable<br>real-time navigation with minimal human intervention. Key components include neural networks<br>trained on camera feeds, ultrasonic sensors, and radar.<\/p>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Examples:<br>o Tesla\u2019s Smart Summon feature autonomously drives cars to owners in parking<br>lots.<br>o AI Day presentations demonstrate the vision-first approach, avoiding LiDAR and<br>focusing on camera-based perception.<br>o Dojo Supercomputer trains its neural networks using millions of driving videos.<\/li>\n\n\n\n<li>Videos and research on Tesla\u2019s capabilities highlight advancements in lane detection,<br>object prediction, and traffic analysis.<br>Exercise 2: Deep Fake Videos<br>Observations:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Best Video: The Barack Obama deep fake video created by Jordan Peele. It convincingly<br>showcased how AI can manipulate speech and facial expressions.<\/li>\n\n\n\n<li>How They Did It: This video was created using GANs (Generative Adversarial<br>Networks) and facial mapping technology. Specific tools like DeepFaceLab or customtrained models were used to replicate Obama\u2019s speech patterns and expressions.<\/li>\n\n\n\n<li>Time and Hardware: Likely required weeks of training on high-end GPUs, given its<br>quality and precision.<br>3<\/li>\n\n\n\n<li>Realism: While highly realistic, close inspection reveals slight lip-sync inconsistencies<br>and unnatural transitions in facial movements.<br>Exercise 3: Wacky Object Variations<br>Category: Potted Plants<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>Found images include:<br>o A plant shaped like a cat.<br>o A pot designed as a robot holding a small cactus.<br>o A plant growing from a shoe instead of a traditional pot.<br>Exercise 4: Rise of Deep Learning<br>Key Points:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Trends in GPU Price Performance:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>GPUs have improved dramatically in computational power while becoming more<br>affordable, enabling broader adoption of deep learning.<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>NVIDIA and AI Dominance:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>NVIDIA\u2019s GPUs revolutionized AI training, surpassing CPU-based methods,<br>with partnerships accelerating AI integration in diverse industries.<br>Exercise 5: Object Detection<\/li>\n\n\n\n<li>Tested images include the unusual plants from Exercise 3.<\/li>\n\n\n\n<li>Observations:<br>o Accurate detection of plant-like shapes.<br>o Unique pots were classified as unrelated objects like furniture.<br>4<br>Exercise 6: Image Classification<br>Observations:<\/li>\n<\/ul>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Single Object: A clean headshot was identified with high confidence and minimal<br>ambiguity.<\/li>\n\n\n\n<li>Cluttered Scene: A coffee cup amidst a desk setup confused the model, leading to<br>multiple low-confidence classifications.<br>Exercise 7: One-Stage vs Two-Stage Models<br>Models:<\/li>\n<\/ol>\n\n\n\n<ul class=\"wp-block-list\">\n<li>YOLO (One-Stage):<br>o Speed: ~45 FPS<br>o Accuracy: ~55 mAP on COCO dataset.<\/li>\n\n\n\n<li>Faster-RCNN (Two-Stage):<br>o Speed: ~5 FPS<br>o Accuracy: ~60-65 mAP on COCO dataset.<br>Comparison: YOLO prioritizes speed for real-time applications, while Faster-RCNN offers<br>higher accuracy at the cost of slower inference.<br>Bonus Insights:<\/li>\n\n\n\n<li>DETR achieves mAP scores around 65-70 on COCO but has slower inference times (~10<br>FPS).<\/li>\n<\/ul>\n\n\n\n<p>Task 1<br>Quantum Computing: Transforming the Digital Frontier<\/p>\n\n\n\n<ol class=\"wp-block-list\">\n<li>Introduction<br>Quantum computing advances with the very principles of quantum mechanics to process<br>information in ways utterly beyond the powers of ordinary computers (Rath et al., 2025). With a<br>promise exponential improvements in performance, quantum computing will revolutionize<br>various industries including cryptography, pharmaceuticals, and artificial intelligence. This latest<br>technology has the potential to refine computational limits while solving hitherto unsolvable<br>problems.<\/li>\n\n\n\n<li>Development History<br>Early Foundations (1980s-1990s)<br>1980: Richard Feynman and Yuri Manin suggest quantum computation as a means to simulate<br>quantum systems.<br>1994: Peter Shor creates Shor\u2019s algorithm that enhances factorization of large numbers,<br>highlighting quantum computing\u2019s potential for cryptography.<br>1998: First experimental implementation of quantum algorithm in 2-qubit quantum computer<br>(Rath et al., 2025).<br>Advancements (2000s-2010s)<br>2001: IBM implements Shor\u2019s algorithm with a 7-qubit quantum computer.<br>2007: Development of quantum annealing to solve optimization problems by D-wave Systems.<br>2019: Google uses Sycamore processor to solve a problem quickly.<br>Recent Progress (2020s-Present)<br>2021: IBM presents a 127-qubit processor \u2013 the Eagle, which expands the scalability limits.<br>3<br>2023: New ways of correcting errors enhance the reliability of quantum computing.<br>2024: Introduction of Hybrid quantum-classical algorithms for solving everyday challenges.<\/li>\n\n\n\n<li>Current Research and Applications<br>Key Research Areas<br>Scalability: Creating quantum processors with an increasingly larger number of qubits, while<br>decreasing their noise and errors.<br>Quantum Algorithms: Development of algorithms for optimization, cryptography, and<br>materials science.<br>Error Correction: Advancing fault-tolerant quantum computing to make systems more robust.<br>Applications<br>Cryptography: Creating quantum-safe encryption systems to counter quantum decryption<br>threats.<br>Drug Discovery: Simulation of molecular interactions for faster pharmaceutical innovations.<br>Logistics and Optimization: Enhnacing supply chain management as well as operational<br>efficiencies.<br>Artificial Intelligence: Leveraging quantum algorithms to improve machine learning models<br>(Rath et al., 2025).<\/li>\n\n\n\n<li>Future Trends<br>Near-Term (5-10 years)<br>i. Scaling quantum processors to hundreds or thousands of qubits.<br>ii. Wide adoption of hybrid quantum-classical systems.<br>iii. Greater industry collaboration to drive more pragmatic quantum solutions.<br>Long-Term (10+ years)<br>4<br>i. Fully fault-tolerant quantum computers able to perform complex simulations.<br>ii. Breakthroughs in material science, AI, and cryptography.<\/li>\n\n\n\n<li>Impacts<br>Societal Impacts<br>Advantages:<br>i. Revolutionizing healthcare through personalized medicine and drug development.<br>ii. Enhance cybersecurity with the use of quantum encryption.<br>iii. Addressing global challenges of climate modeling and sustainable energy solutions.<br>Disadvantages:<br>i. Possible misuse in breaking existing symmetric encryption standards.<br>ii. High development costs, greatly reducing accessibility.<br>iii. Ethical concerns over the disruptive impact to industries and jobs (Rath et al., 2025).<br>Industrial Impacts<br>Advantages:<br>i. Transforming various industries including but not limited to finance, logistics, and<br>pharmaceuticals.<br>ii. Accelerating the pace of innovation in materials and manufacturing (Rath et al.,<br>2025).<br>Disadvantages:<br>i. Quantum infrastructure requires huge investments.<br>ii. Risks associated with emerging technologies&#8217; dependence.<br>Impact on Daily Life<br>Advantages:<br>5<br>i. Improvement in cybersecurity for the protection of personal data.<br>ii. Possible breakthroughs in personalized healthcare (Ravindran et al., 2025).<br>Disadvantages:<br>i. Privacy concerns regarding security of data.<br>ii. Widening digital gap because of unequal access to quantum resources (Ravindran et<br>al., 2025).<br>Conclusion<br>Quantum computing is a revolutionary change in technology and holds immense promise for<br>solving complicated problems in various walks of life. Though scalability and error correction<br>are serious challenges, improvements being made are certain to lead toward practical<br>applications. But great potential means great responsibility too, so ethics in development and<br>equality in access will be very crucial in leveraging its benefits with minimal risks.<br>6<br>Task 2<br>Remote Healthcare Solution Plan<br>7<br>8<br>Figure 1.0: Flowchart<br>Overview<br>The solution will leverage the emerging technologies to facilitate remote health delivery, secure<br>payment, and innovative delivery of medication to a patient in London who is consulting a<br>doctor in Sunderland.<\/li>\n\n\n\n<li>Teleconsultation Solution<br>Primary Technology: Secure Telemedicine Platform<br>Implementation:<br>o Video conferencing system compliant with HIPAA<br>o Integration with EHR<br>o Secure document sharing and generation of digital prescription<br>o Integrated appointment scheduling and reminder system<br>Security Measures:<br>o End-to-end encryption for all communications<br>o Multi-factor authentication for both patient and doctor<br>o Secure storage of medical records and consultation history<\/li>\n\n\n\n<li>Secure Payment System<br>Technology Implementation:<br>o Integration with a secure payment gateway that supports multiple payment methods<br>o Blockchain-based transaction system for enhanced security<br>o Smart contract that allows for automatic payment processing<br>o Real-time payment confirmation system<br>Security Features:<br>9<br>o Tokenization of payment information<br>o PCI DSS compliance<br>o Transaction monitoring and fraud detection<br>o Encrypted payment data transmission<\/li>\n\n\n\n<li>Innovative Medication Delivery<br>Primary Solution: Drone Delivery System<br>o Autonomous drone delivery network between pharmacy and patient<br>o Real-time tracking and monitoring system<br>o Weather-resistant drone design for reliable service<br>o Secure storage compartment with temperature control<br>Backup Solutions:<br>o Partnership with local courier services for last-mile delivery<br>o Option for autonomous ground robots in case of drone restrictions<br>o Emergency delivery protocol for urgent medications<br>Technical Requirements<br>Infrastructure Needs:<br>o High-speed internet connectivity<br>o Cloud-based system architecture<br>o Backup servers for continuous service<br>o Mobile application support<br>Integration Points:<br>o Hospital management system<br>o Pharmacy inventory system<br>10<br>o Payment processing system<br>o Delivery management system<br>Safety and Compliance<br>Regulatory Compliance:<br>o NHS Digital standards compliance<br>o GDPR compliance for data protection<br>o CAA regulations for drone operations<br>o Electronic Prescription Service (EPS) integration<br>Quality Assurance:<br>o Regular system audits<br>o Performance monitoring<br>o User feedback collection<br>o Continuous improvement protocol<br>11<br>References<br>Rath, K.C., Khang, A., Mohanta, G.K., Panda, R.A. and Sahu, R. (2025). The Quantum Shift:<br>Transformative Innovations in the Digital Realm. In The Quantum Evolution (pp. 1-26). CRC<br>Press.<br>Ravindran, D., Revathi, S., Sowndharya, V., Farzhana, I., Sathya, V., Girija, P. and<br>Subramanian, S. (2025). Unraveling the Quantum Computing Frontier: Advancements,<br>Challenges, and Future Prospects. Integration of AI, Quantum Computing, and Semiconductor<br>Technology, pp.139-158.<\/li>\n<\/ol>\n\n\n<p><\/div><\/p>\n","protected":false},"excerpt":{"rendered":"","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[359],"tags":[],"class_list":["post-37705","post","type-post","status-publish","format-standard","hentry","category-education"],"_links":{"self":[{"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/posts\/37705","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/comments?post=37705"}],"version-history":[{"count":1,"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/posts\/37705\/revisions"}],"predecessor-version":[{"id":37706,"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/posts\/37705\/revisions\/37706"}],"wp:attachment":[{"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/media?parent=37705"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/categories?post=37705"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/myassignmenthelp.info\/assignments\/wp-json\/wp\/v2\/tags?post=37705"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}