Mycelium Robotics

Perception Engineer Interview Questions: What to Ask and What to Look For

Published April 2026 · Mycelium

Last updated: April 2026

Perception interviews fail when they test ML theory without testing systems thinking. The strongest perception engineers understand sensor physics, real-time constraints, and failure modes, not just model architectures. A candidate who can explain how to handle LiDAR returns in heavy rain is more valuable than one who can recite the latest detection paper.

Most interview loops over-index on model training and under-index on deployment. The result is a team that can publish papers but cannot ship a reliable perception stack. The questions in this guide are designed to separate engineers who have built and deployed perception systems from those who have only trained models on clean datasets.

Whether you are hiring for a perception engineer role at an autonomous vehicle company or a warehouse robotics startup, the fundamentals are the same. You need someone who thinks about the full pipeline from raw sensor data to actionable output, understands the perception market landscape, and can reason about failure modes before they become field incidents.

Screening questions

These questions work well in a 30-minute phone screen. They quickly reveal whether a candidate has hands-on experience with real perception systems or has only worked in academic settings. Listen for specificity. Strong candidates name sensors, frame rates, latency budgets, and deployment environments without being prompted.

Q: “Walk me through a perception pipeline you built or contributed to. What sensors were involved, and what was the output?”

Strong answer: Describes specific sensors (e.g., Velodyne VLP-16, 4x GMSL cameras), the fusion approach used, the latency budget they operated within, and the deployment context. Mentions concrete outputs like 3D bounding boxes at 10Hz feeding into a planner, or semantic segmentation masks consumed by a navigation stack. Explains their specific contribution versus team effort.

Red flags: Only describes training a model on a public dataset. No mention of real-time constraints or deployment. Cannot distinguish their contribution from the team's work. Uses vague language like “we used deep learning for detection.”

Q: “What is the difference between early fusion and late fusion in multi-sensor systems? When would you choose each?”

Strong answer: Explains that early fusion combines raw sensor data before processing (e.g., projecting LiDAR points onto camera images), while late fusion processes each sensor independently and merges results at the object level. Discusses tradeoffs: early fusion captures richer correlations but couples sensor modalities and increases latency; late fusion is more modular and resilient to single-sensor failure but may miss cross-modal information. Provides a real example of choosing one approach over the other.

Red flags: Gives a textbook definition without practical experience. Cannot explain when one approach is preferable. Has never built a multi-sensor system.

Q: “How do you evaluate a detection model beyond mAP? What metrics matter for deployed systems?”

Strong answer: Mentions inference latency, false positive rate at the specific operating point (not just the AUC), performance degradation in edge cases (rain, dust, low light), temporal consistency of detections across frames, recall at safety-critical IoU thresholds, and power consumption on target hardware. Understands that mAP hides critical failure modes because it averages across classes and thresholds.

Red flags: Only mentions mAP, precision, and recall. Has never thought about metrics in the context of a downstream consumer like a planner. Does not consider latency as a metric.

Q: “Describe a failure mode you encountered in a deployed perception system. How did you debug it?”

Strong answer: Tells a specific story with a clear root cause and resolution. For example: “Our LiDAR-based pedestrian detector was producing false positives on chain-link fences at a specific distance range. We traced it to the point cloud density at 30-40m creating patterns similar to leg clusters. We added a geometric consistency check and retrained with hard negatives from those scenes.” Shows systematic debugging, not guesswork.

Red flags: Cannot describe a real failure from their own experience. Gives a hypothetical answer. Says they “just retrained with more data” without analyzing the root cause.

Q: “What are the key differences between working with LiDAR point clouds versus camera images for 3D detection?”

Strong answer: Discusses that LiDAR provides direct depth measurements but is sparse (especially at range), while cameras offer dense texture and color information but require depth estimation. Covers different network architectures (PointNet/PointPillars for point clouds vs. image backbones with depth heads), calibration challenges between modalities, range and resolution tradeoffs, and cost implications. Mentions that LiDAR degrades in fog and dust while cameras struggle in low light and direct sun.

Red flags: Has only worked with one modality and cannot meaningfully compare. Does not understand the fundamental differences in data representation. Cannot name specific architectures for either modality.

Q: “What is your experience with model optimization for edge deployment? What techniques have you used?”

Strong answer: Discusses specific techniques like quantization (INT8, FP16), pruning, knowledge distillation, TensorRT optimization, or ONNX conversion. Can explain the accuracy-latency tradeoffs of each approach and has measured the impact on their specific models. Knows which layers are most sensitive to quantization and how to use calibration datasets.

Red flags: Has only trained models in the cloud and never deployed to edge hardware. Cannot explain the difference between FP32, FP16, and INT8 inference. Thinks model optimization means hyperparameter tuning.

Technical deep dive questions

These questions belong in a 60-minute technical round led by someone with perception domain experience. They test the ability to reason about real systems under constraints. A general software engineer cannot properly evaluate answers to these questions. If you do not have a perception expert on your interview panel, you are likely making bad hiring decisions.

Q: “You are building a sensor fusion pipeline with one LiDAR and four cameras. How do you handle temporal misalignment between sensors?”

Strong answer: Discusses hardware-level synchronization using PTP (Precision Time Protocol) or trigger signals to minimize temporal offset at the source. Explains that residual misalignment requires compensating for ego-motion between sensor timestamps using IMU data or odometry. Describes interpolation strategies for aligning sensor data to a common reference time. Mentions that extrinsic calibration must account for rolling shutter effects on cameras. Understands that even 10ms of misalignment at highway speeds creates 30cm of spatial error.

Red flags: Assumes sensors are synchronized by default. Does not understand PTP or hardware triggering. Cannot quantify the impact of temporal misalignment on downstream accuracy.

Q: “Your 3D detection model performs well on test data but fails on rainy night scenes. Walk me through your debugging process.”

Strong answer: Starts with data analysis, not model changes. Checks whether rainy night scenes are underrepresented in training data. Examines raw sensor data to understand how rain affects each modality: LiDAR returns scatter off raindrops creating noise, camera images suffer from glare and reduced contrast. Investigates whether the preprocessing pipeline handles these degraded inputs. Looks at specific failure examples to categorize error types. Only then considers targeted solutions: domain-specific augmentation, sensor-specific preprocessing (rain filtering for LiDAR), model architecture changes, or additional training data from those conditions.

Red flags: Immediately jumps to “collect more rainy night data” or “retrain the model.” Does not consider sensor-level effects. Has no systematic debugging methodology. Treats it as purely a data problem.

Q: “How would you design an object tracking system that maintains identity through occlusions?”

Strong answer: Describes a multi-stage approach: a prediction model (Kalman filter or learned predictor) that maintains estimated state during occlusion, re-identification features (appearance, shape, motion pattern) for matching when objects reappear, and track management logic for handling birth, death, and resurrection of tracks. Discusses uncertainty growth during occlusion and how to set thresholds for when to drop a track versus keep predicting. Mentions handling partial occlusions differently from full occlusions. May reference specific algorithms like SORT, DeepSORT, or more recent transformer-based trackers.

Red flags: Treats each frame independently without temporal reasoning. Cannot explain how to maintain track identity. Does not consider uncertainty growth during occlusion. Only knows detection, not tracking.

Q: “Explain how you would calibrate a LiDAR-camera system in the field. What happens when calibration drifts?”

Strong answer: Describes target-based methods (checkerboard patterns visible in both modalities) for initial calibration and targetless methods (edge alignment, mutual information) for refinement. Explains that calibration drifts due to thermal expansion, vibration, and mechanical shock. Discusses online calibration approaches that continuously estimate and correct extrinsic parameters using natural features in the environment. Mentions monitoring metrics like projection error or feature alignment scores to detect drift before it causes downstream failures. Knows that a 0.5-degree calibration error at 50m creates a 40cm projection offset.

Red flags: Treats calibration as a one-time setup procedure. Does not know how to detect calibration drift. Cannot explain targetless calibration methods. Has no awareness of thermal or mechanical effects on sensor mounting.

Q: “What are the tradeoffs between running perception on a GPU versus a dedicated inference accelerator like a Jetson Orin or edge TPU?”

Strong answer: Discusses latency (dedicated accelerators often provide more deterministic inference times), power consumption (critical for battery-powered robots), model compatibility (not all operations are supported on all accelerators), quantization requirements (many accelerators require INT8), development iteration speed (GPU is faster for prototyping), and cost at volume. Understands that TensorRT on NVIDIA hardware gives good GPU utilization but may require model-specific tuning. Knows that edge TPUs are power-efficient but restrict model architectures to supported operations.

Red flags: Only familiar with one deployment target. Does not consider power consumption. Cannot explain quantization effects on model accuracy. Thinks inference hardware is interchangeable.

Q: “How do you handle the long tail of edge cases in a perception system? You cannot collect data for every scenario.”

Strong answer: Outlines a multi-pronged strategy: simulation and synthetic data generation for rare scenarios, domain randomization to improve generalization, active learning to prioritize annotation of informative examples from field data, hard example mining from logged data, safety margins in the system design so perception failures do not immediately cause unsafe behavior, and graceful degradation strategies (e.g., slowing down when detection confidence is low). Understands that this is a system-level problem, not just a data problem.

Red flags: Believes that collecting more data always solves the problem. Has no strategy for synthetic data or simulation. Does not consider system-level safety margins. Thinks the long tail can be eliminated with enough training data.

Q: “Describe how you would implement ground plane estimation from a LiDAR point cloud. Why does this matter for object detection?”

Strong answer: Describes using RANSAC or a similar robust estimator to fit a plane to the dominant ground surface. Explains that real-world ground is not perfectly flat, so the estimator needs to handle slopes, curbs, and speed bumps, potentially using a piecewise planar model or elevation grid. Ground plane estimation matters because removing ground points dramatically reduces the search space for object detection, enables height-based filtering to separate objects from ground clutter, and provides a reference for estimating object height and position. Also useful for detecting drivable surface boundaries.

Red flags: Cannot explain why ground removal matters for object detection. Does not know RANSAC or any robust estimation method. Assumes the ground is always flat. Has never worked directly with point cloud data.

Q: “How would you design a perception system that degrades gracefully when a sensor fails mid-operation?”

Strong answer: Discusses sensor health monitoring to detect failures quickly, fallback perception modes that operate on reduced sensor sets, communicating confidence degradation to downstream systems (planner should know that perception is operating in a degraded state), reducing operational speed or scope when sensor redundancy is lost, and logging the failure for post-incident analysis. May describe specific examples like switching from LiDAR-camera fusion to camera-only detection with wider safety margins when LiDAR fails.

Red flags: Has never considered sensor failure in their design. Assumes all sensors are always available. Does not understand the concept of graceful degradation. Would halt the robot entirely rather than operating in a reduced capacity.

System design questions

System design rounds reveal how a candidate thinks at the architecture level. These are open-ended by design. A strong candidate will ask clarifying questions, make explicit tradeoffs, and acknowledge uncertainty. Give them a whiteboard (or virtual equivalent) and 45 to 60 minutes.

Q: “Design a perception system for an autonomous forklift operating in a mixed warehouse with humans, other forklifts, and dynamic inventory.”

Strong answer: Starts by asking about the operating environment: aisle width, lighting conditions, ceiling height, speed requirements. Selects sensors with rationale (e.g., 3D LiDAR for obstacle detection in aisles, cameras for pallet identification and human detection, ultrasonic sensors for close-range safety). Addresses the safety-critical requirement of human detection with explicit recall targets. Distinguishes between static map elements (shelving, walls) and dynamic objects (humans, other forklifts, pallets being moved). Discusses latency requirements for safe stopping distances at operating speed. Considers failure modes: blocked sensors from dust, sensor occlusion in narrow aisles, lighting changes between indoor and loading dock areas.

Red flags: Jumps to sensor selection without understanding requirements. Does not prioritize human safety detection. Ignores the warehouse-specific challenges like dust, narrow aisles, and dynamic inventory. Proposes an outdoor AV perception stack without adapting to the indoor warehouse context.

Q: “You are architecting the perception stack for a new robot platform. The team wants to use a single end-to-end neural network for all perception. Make the case for or against this approach.”

Strong answer: Presents both sides with genuine understanding. For: reduces integration complexity, allows the network to learn shared representations across tasks, avoids cascading errors between modules, and end-to-end optimization can find efficiencies that modular systems miss. Against: debugging is extremely difficult because you cannot isolate which perception task is failing, data requirements are massive, failure in one task can corrupt others, updating one capability requires retraining and revalidating everything, and safety certification is harder for a monolithic system. A strong candidate takes a clear position but acknowledges the legitimate arguments on the other side. The best candidates discuss hybrid approaches that capture benefits of both.

Red flags: Dogmatically supports one approach without considering the other. Cannot articulate the debugging and validation challenges of end-to-end systems. Does not consider safety certification implications. Has strong opinions but no experience with either approach.

Q: “Your company is scaling from 10 robots to 1,000 robots deployed across different geographies. How does this change your perception system design?”

Strong answer: Addresses data pipeline scaling (collecting, labeling, and managing training data from a large fleet), model update and deployment mechanisms (OTA updates, A/B testing, rollback), geographic variation in environments and edge cases, monitoring and alerting for perception performance across the fleet, and hardware consistency challenges across manufacturing batches. Discusses how fleet data can be leveraged for active learning and continuous improvement. Considers regulatory differences across geographies.

Red flags: Thinks scaling is just deploying the same model to more robots. Does not consider fleet-level data management. Has no experience with model deployment at scale. Ignores geographic and environmental variation.

Culture and collaboration questions

Perception engineers do not work in isolation. They ship to planners, take requirements from systems engineers, and depend on data infrastructure teams. These questions test whether a candidate can operate effectively in a robotics organization where cross-team alignment determines product velocity.

Q: “How do you work with the planning team when your perception system has uncertainty about an object classification?”

Strong answer: Describes passing confidence scores and uncertainty estimates to the planner rather than just hard classifications. Explains how they have worked with planning teams to define interfaces, agree on uncertainty representations, and set thresholds collaboratively. Understands that the planner needs to make different decisions based on whether an uncertain detection is near the robot's path or far away.

Red flags: Treats perception as an isolated component. Only outputs hard classifications with no uncertainty information. Has never collaborated with a downstream consumer of their perception outputs.

Q: “A field deployment is tomorrow and your perception system has a known failure case. What do you do?”

Strong answer: Immediately communicates the issue to stakeholders with a clear description of the failure mode, its severity, and the conditions that trigger it. Proposes mitigations: can the operational domain be restricted to avoid the triggering condition? Can a runtime check detect and handle the failure case? Is a human operator available as a fallback? Makes a clear recommendation about whether to proceed with mitigations or delay. Documents everything.

Red flags: Stays silent and hopes the failure case does not occur. Pushes the decision entirely to management without providing a technical assessment. Refuses to deploy under any circumstances without considering mitigations. Does not document the known issue.

Q: “How do you prioritize perception work when the autonomy team, controls team, and product team all have competing requests?”

Strong answer: Describes a framework for prioritization that considers safety impact, deployment timeline, effort required, and dependencies. Has experience negotiating with multiple stakeholders and making tradeoffs explicit. Does not simply default to whoever has the loudest voice or the highest title. Can give a concrete example of a time they had to say no or negotiate scope.

Red flags: Says yes to everything. Has no prioritization framework. Cannot push back on unreasonable requests. Does not understand how their work fits into the broader product roadmap.

Q: “Tell me about a time you disagreed with a technical decision on your team. How did you handle it?”

Strong answer: Describes a specific disagreement with technical substance. Explains how they gathered data or built a prototype to support their position. Shows willingness to commit to the team's decision even if it was not their preferred approach. Demonstrates that they can disagree without being disagreeable.

Red flags: Claims they have never disagreed with anyone. Describes a disagreement where they were clearly right and everyone else was wrong. Cannot describe how they resolved the disagreement constructively.

Recommended interview process

We recommend a four-stage interview loop for perception engineers:

Stage 1: Phone screen (30 minutes). Use the screening questions above. The goal is to confirm that the candidate has real perception experience and can articulate it clearly. A hiring manager or senior perception engineer should run this round.

Stage 2: Perception-specific technical (60 minutes). This round must be led by someone with deep perception experience. Use the technical deep dive questions. Include at least one debugging scenario and one design question. Do not substitute a generic coding interview for this round. You are testing perception systems thinking, not LeetCode ability.

Stage 3: System design (60 minutes). Use one of the system design questions above or adapt one to your specific domain. Give the candidate time to ask clarifying questions and iterate on their design. The best signal comes from how they handle ambiguity and tradeoffs, not whether they arrive at a specific answer.

Stage 4: Culture and collaboration (45 minutes). Use the culture questions above, supplemented with questions specific to your team dynamics. This round is best conducted by a cross-functional partner (e.g., a planning or controls engineer who would work closely with this hire).

The most important detail: the technical round must be led by someone with perception domain expertise. A generalist software engineer will not know whether an answer about sensor fusion or calibration is correct. If you do not have this expertise internally, consider bringing in an external advisor for the interview panel. For a deeper look at structuring the overall process, see our guide on how to hire a robotics engineer.

For compensation benchmarking to help close candidates, refer to our San Francisco robotics salary guideor the equivalent guide for your hiring location. Perception engineers in the Bay Area command $160k to $250k+ base depending on seniority, with significant equity upside at growth-stage companies. Being unprepared on compensation wastes everyone's time.

If you need help building a perception engineering team, our search services are designed specifically for robotics organizations.