Udvidet returret til d. 31. januar 2025

Bøger af Kenichi Kanatani

Filter
Filter
Sorter efterSorter Populære
  • af Margrit Betke
    657,95 kr.

    Because circular objects are projected to ellipses in images, ellipse fitting is a first step for 3-D analysis of circular objects in computer vision applications. For this reason, the study of ellipse fitting began as soon as computers came into use for image analysis in the 1970s, but it is only recently that optimal computation techniques based on the statistical properties of noise were established. These include renormalization (1993), which was then improved as FNS (2000) and HEIV (2000). Later, further improvements, called hyperaccurate correction (2006), HyperLS (2009), and hyper-renormalization (2012), were presented. Today, these are regarded as the most accurate fitting methods among all known techniques. This book describes these algorithms as well implementation details and applications to 3-D scene analysis. We also present general mathematical theories of statistical optimization underlying all ellipse fitting algorithms, including rigorous covariance and bias analyses and the theoretical accuracy limit. The results can be directly applied to other computer vision tasks including computing fundamental matrices and homographies between images. This book can serve not simply as a reference of ellipse fitting algorithms for researchers, but also as learning material for beginners who want to start computer vision research. The sample program codes are downloadable from the website: https://sites.google.com/a/morganclaypool.com/ellipse-fitting-for-computer-vision-implementation-and-applications.

  • af Kenichi Kanatani
    570,95 kr.

    Modeling data from visual and linguistic modalities together creates opportunities for better understanding of both, and supports many useful applications. Examples of dual visual-linguistic data includes images with keywords, video with narrative, and figures in documents. We consider two key task-driven themes: translating from one modality to another (e.g., inferring annotations for images) and understanding the data using all modalities, where one modality can help disambiguate information in another. The multiple modalities can either be essentially semantically redundant (e.g., keywords provided by a person looking at the image), or largely complementary (e.g., meta data such as the camera used). Redundancy and complementarity are two endpoints of a scale, and we observe that good performance on translation requires some redundancy, and that joint inference is most useful where some information is complementary. Computational methods discussed are broadly organized into ones for simple keywords, ones going beyond keywords toward natural language, and ones considering sequential aspects of natural language. Methods for keywords are further organized based on localization of semantics, going from words about the scene taken as whole, to words that apply to specific parts of the scene, to relationships between parts. Methods going beyond keywords are organized by the linguistic roles that are learned, exploited, or generated. These include proper nouns, adjectives, spatial and comparative prepositions, and verbs. More recent developments in dealing with sequential structure include automated captioning of scenes and video, alignment of video and text, and automated answering of questions about scenes depicted in images.

  • af Kenichi Kanatani
    610,95 kr.

    Linear algebra is one of the most basic foundations of a wide range of scientific domains, and most textbooks of linear algebra are written by mathematicians. However, this book is specifically intended to students and researchers of pattern information processing, analyzing signals such as images and exploring computer vision and computer graphics applications. The author himself is a researcher of this domain. Such pattern information processing deals with a large amount of data, which are represented by high-dimensional vectors and matrices. There, the role of linear algebra is not merely numerical computation of large-scale vectors and matrices. In fact, data processing is usually accompanied with "e;geometric interpretation."e; For example, we can think of one data set being "e;orthogonal"e; to another and define a "e;distance"e; between them or invoke geometric relationships such as "e;projecting"e; some data onto some space. Such geometric concepts not only help us mentally visualize abstract high-dimensional spaces in intuitive terms but also lead us to find what kind of processing is appropriate for what kind of goals. First, we take up the concept of "e;projection"e; of linear spaces and describe "e;spectral decomposition,"e; "e;singular value decomposition,"e; and "e;pseudoinverse"e; in terms of projection. As their applications, we discuss least-squares solutions of simultaneous linear equations and covariance matrices of probability distributions of vector random variables that are not necessarily positive definite. We also discuss fitting subspaces to point data and factorizing matrices in high dimensions in relation to motion image analysis. Finally, we introduce a computer vision application of reconstructing the 3D location of a point from three camera views to illustrate the role of linear algebra in dealing with data with noise. This book is expected to help students and researchers of pattern information processing deepen the geometric understanding of linear algebra.

  • - Parameter Computation and Lie Algebra based Optimization
    af Kenichi Kanatani
    612,95 - 1.104,95 kr.

  • - Geometric Analysis and Implementation
    af Kenichi Kanatani, Yasuyuki Sugaya & Yasushi Kanazawa
    439,95 - 693,95 kr.

    Unlike other computer vision textbooks, this guide takes a unique approach in which the initial focus is on practical application and the procedures necessary to actually build a computer vision system.

  • - Theory and Practice
    af Kenichi Kanatani
    287,95 kr.

    This text discusses the mathematical foundations of statistical inference for building 3-dimensional models from image and sensor data that contain noise -- a task involving autonomous robots guided by video cameras and sensors. The text employs a theoretical accuracy for the optimization procedure, which maximizes the reliability of estimations based on noise data. 1996 edition.

Gør som tusindvis af andre bogelskere

Tilmeld dig nyhedsbrevet og få gode tilbud og inspiration til din næste læsning.