Buy Regularized Estimation of Covariance Matrices for Longitudinal Data Through Smoothing and Shrinkage
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Book 1
Book 2
Book 3
Home > Art, Film & Photography > Architecture > Regularized Estimation of Covariance Matrices for Longitudinal Data Through Smoothing and Shrinkage
Regularized Estimation of Covariance Matrices for Longitudinal Data Through Smoothing and Shrinkage

Regularized Estimation of Covariance Matrices for Longitudinal Data Through Smoothing and Shrinkage


     0     
5
4
3
2
1



Out of Stock


Notify me when this book is in stock
X
About the Book

This dissertation develops methods that efficiently estimate a covariance matrix of longitudinal data without making restrictive parametric assumptions, even when the dimension is large relative to the sample size. The high dimensionality is an important issue because the number of free elements in the covariance matrix increases quadratically with its dimension. Charles Stein pointed out in the 1950's that the unrestricted maximum likelihood estimator of the covariance matrix is statistically inefficient. To gain statistical efficiency and overcome the high dimensionality problem in covariance matrix estimation, an effective regularization scheme is needed. In this dissertation we design such schemes using the regression formulation of covariance matrix estimation advocated by Pourahmadi (1999, 2000). We propose penalized likelihood methods for producing statistically efficient estimator of a covariance matrix for longitudinal data. The approaches parameterize the covariance matrix through the modified Cholesky decomposition. For longitudinal data, the entries of the lower triangular and the diagonal matrix associated with the modified Cholesky decomposition can be interpreted as regression coefficients and prediction error variances when regressing a measurement on its predecessors. Many covariance matrix estimation methods have been developed based on two observations: first, there is usually some kind of continuity among neighboring elements in the lower triangular; second, the lower triangular is likely to have many off-diagonal elements that are zero or close to zero. Wu and Pourahmadi (2003) and Huang, Liu, and Liu (2007) proposed to smooth the long subdiagonals of the lower triangular using local polynomial or splines techniques while truncate the short subdiagonals to zeros. Their approaches are rather restrictive and can't be easily combined with shrinkage. Huang et al. (2006) proposed to shrink the elements of the lower triangular by imposing the L1 or L2 penalty using the penalized likelihood, but it completely ignores the possible smoothness in the modified Cholesky factor. Our first new approach relaxed these restrictions and employs roughness penalty to impose smoothness in the rows or subdiagonals of the lower triangular. Use of roughness has been well studied for nonparametric function estimation and the second-order roughness penalty can be viewed as an approximation of the integrated squared second derivative penalty. This proposed new smoothing method can be easily combined with the shrinkage. Our second new approach combines smoothing with shrinkage using penalized likelihood. Our method is general enough to include various combinations of the shrinkage penalty (no penalty, the L1 penalty, or the L2 penalty) and the roughness penalty (no smoothing, smoothing along the rows of the lower triangular matrix, or smoothing along the subdiagonals of the lower triangular matrix), while the best combination can be chosen using a data-driven method. It turns out that combination of shrinkage and smoothing does better and sometimes much better than using shrinkage or smoothing alone. Our simulation study shows the superior performance of this new method. Most of existing works on covariance matrix estimation are only directly applicable to balanced longitudinal data. In this dissertation we also consider the unbalanced case by treating it as a missing data problem and the EM algorithm can be easily applied to our proposed approach. The practical value of our methods is demonstrated in its application to efficient estimation of regression mean parameters for longitudinal data. It is well known that the regression mean...


Best Sellers


Product Details
  • ISBN-13: 9781244076815
  • Publisher: Proquest, Umi Dissertation Publishing
  • Publisher Imprint: Proquest, Umi Dissertation Publishing
  • Height: 246 mm
  • Weight: 186 gr
  • ISBN-10: 1244076813
  • Publisher Date: 01 Sep 2011
  • Binding: Paperback
  • Spine Width: 5 mm
  • Width: 189 mm

Related Categories

Similar Products

Add Photo
Add Photo

Customer Reviews

REVIEWS      0     
Click Here To Be The First to Review this Product
Regularized Estimation of Covariance Matrices for Longitudinal Data Through Smoothing and Shrinkage
Proquest, Umi Dissertation Publishing -
Regularized Estimation of Covariance Matrices for Longitudinal Data Through Smoothing and Shrinkage
Writing guidlines
We want to publish your review, so please:
  • keep your review on the product. Review's that defame author's character will be rejected.
  • Keep your review focused on the product.
  • Avoid writing about customer service. contact us instead if you have issue requiring immediate attention.
  • Refrain from mentioning competitors or the specific price you paid for the product.
  • Do not include any personally identifiable information, such as full names.

Regularized Estimation of Covariance Matrices for Longitudinal Data Through Smoothing and Shrinkage

Required fields are marked with *

Review Title*
Review
    Add Photo Add up to 6 photos
    Would you recommend this product to a friend?
    Tag this Book Read more
    Does your review contain spoilers?
    What type of reader best describes you?
    I agree to the terms & conditions
    You may receive emails regarding this submission. Any emails will include the ability to opt-out of future communications.

    CUSTOMER RATINGS AND REVIEWS AND QUESTIONS AND ANSWERS TERMS OF USE

    These Terms of Use govern your conduct associated with the Customer Ratings and Reviews and/or Questions and Answers service offered by Bookswagon (the "CRR Service").


    By submitting any content to Bookswagon, you guarantee that:
    • You are the sole author and owner of the intellectual property rights in the content;
    • All "moral rights" that you may have in such content have been voluntarily waived by you;
    • All content that you post is accurate;
    • You are at least 13 years old;
    • Use of the content you supply does not violate these Terms of Use and will not cause injury to any person or entity.
    You further agree that you may not submit any content:
    • That is known by you to be false, inaccurate or misleading;
    • That infringes any third party's copyright, patent, trademark, trade secret or other proprietary rights or rights of publicity or privacy;
    • That violates any law, statute, ordinance or regulation (including, but not limited to, those governing, consumer protection, unfair competition, anti-discrimination or false advertising);
    • That is, or may reasonably be considered to be, defamatory, libelous, hateful, racially or religiously biased or offensive, unlawfully threatening or unlawfully harassing to any individual, partnership or corporation;
    • For which you were compensated or granted any consideration by any unapproved third party;
    • That includes any information that references other websites, addresses, email addresses, contact information or phone numbers;
    • That contains any computer viruses, worms or other potentially damaging computer programs or files.
    You agree to indemnify and hold Bookswagon (and its officers, directors, agents, subsidiaries, joint ventures, employees and third-party service providers, including but not limited to Bazaarvoice, Inc.), harmless from all claims, demands, and damages (actual and consequential) of every kind and nature, known and unknown including reasonable attorneys' fees, arising out of a breach of your representations and warranties set forth above, or your violation of any law or the rights of a third party.


    For any content that you submit, you grant Bookswagon a perpetual, irrevocable, royalty-free, transferable right and license to use, copy, modify, delete in its entirety, adapt, publish, translate, create derivative works from and/or sell, transfer, and/or distribute such content and/or incorporate such content into any form, medium or technology throughout the world without compensation to you. Additionally,  Bookswagon may transfer or share any personal information that you submit with its third-party service providers, including but not limited to Bazaarvoice, Inc. in accordance with  Privacy Policy


    All content that you submit may be used at Bookswagon's sole discretion. Bookswagon reserves the right to change, condense, withhold publication, remove or delete any content on Bookswagon's website that Bookswagon deems, in its sole discretion, to violate the content guidelines or any other provision of these Terms of Use.  Bookswagon does not guarantee that you will have any recourse through Bookswagon to edit or delete any content you have submitted. Ratings and written comments are generally posted within two to four business days. However, Bookswagon reserves the right to remove or to refuse to post any submission to the extent authorized by law. You acknowledge that you, not Bookswagon, are responsible for the contents of your submission. None of the content that you submit shall be subject to any obligation of confidence on the part of Bookswagon, its agents, subsidiaries, affiliates, partners or third party service providers (including but not limited to Bazaarvoice, Inc.)and their respective directors, officers and employees.

    Accept

    Fresh on the Shelf


    Inspired by your browsing history


    Your review has been submitted!

    You've already reviewed this product!