@@ -832,53 +832,6 @@ std::tuple<Tensor,Tensor> _th_gels(const Tensor & self, const Tensor & A) {
832
832
}
833
833
return std::tuple<Tensor, Tensor>(res1, res2);
834
834
}
835
- std::tuple<Tensor &,Tensor &> _th_eig_out (Tensor & res1, Tensor & res2, const Tensor & self, bool eigenvectors) {
836
- // DeviceGuard omitted
837
- auto dispatch_scalar_type = infer_scalar_type (self);
838
-
839
- switch (dispatch_scalar_type) {
840
- case ScalarType::Double: {
841
- auto res1_ = checked_dense_tensor_unwrap (res1, " res1" , 0 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
842
- auto res2_ = checked_dense_tensor_unwrap (res2, " res2" , 0 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
843
- auto self_ = checked_dense_tensor_unwrap (self, " self" , 1 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
844
- THDoubleTensor_geev (res1_, res2_, self_, eigenvectors);
845
- break ;
846
- }
847
- case ScalarType::Float: {
848
- auto res1_ = checked_dense_tensor_unwrap (res1, " res1" , 0 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
849
- auto res2_ = checked_dense_tensor_unwrap (res2, " res2" , 0 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
850
- auto self_ = checked_dense_tensor_unwrap (self, " self" , 1 , " _th_eig_out" , false , DeviceType::CPU, dispatch_scalar_type);
851
- THFloatTensor_geev (res1_, res2_, self_, eigenvectors);
852
- break ;
853
- }
854
- default :
855
- AT_ERROR (" _th_eig_out not supported on CPUType for " , dispatch_scalar_type);
856
- }
857
- return std::tuple<Tensor &, Tensor &>(res1, res2);
858
- }
859
- std::tuple<Tensor,Tensor> _th_eig (const Tensor & self, bool eigenvectors) {
860
- // DeviceGuard omitted
861
- auto dispatch_scalar_type = infer_scalar_type (self);
862
- auto res1_ = c10::make_intrusive<TensorImpl, UndefinedTensorImpl>(c10::Storage (c10::Storage::use_byte_size_t (), 0 , allocator (), true ),DispatchKey::CPU, scalarTypeToTypeMeta (dispatch_scalar_type)).release ();
863
- auto res1 = Tensor (c10::intrusive_ptr<TensorImpl, UndefinedTensorImpl>::reclaim (res1_));
864
- auto res2_ = c10::make_intrusive<TensorImpl, UndefinedTensorImpl>(c10::Storage (c10::Storage::use_byte_size_t (), 0 , allocator (), true ),DispatchKey::CPU, scalarTypeToTypeMeta (dispatch_scalar_type)).release ();
865
- auto res2 = Tensor (c10::intrusive_ptr<TensorImpl, UndefinedTensorImpl>::reclaim (res2_));
866
- switch (dispatch_scalar_type) {
867
- case ScalarType::Double: {
868
- auto self_ = checked_dense_tensor_unwrap (self, " self" , 1 , " _th_eig" , false , DeviceType::CPU, dispatch_scalar_type);
869
- THDoubleTensor_geev (res1_, res2_, self_, eigenvectors);
870
- break ;
871
- }
872
- case ScalarType::Float: {
873
- auto self_ = checked_dense_tensor_unwrap (self, " self" , 1 , " _th_eig" , false , DeviceType::CPU, dispatch_scalar_type);
874
- THFloatTensor_geev (res1_, res2_, self_, eigenvectors);
875
- break ;
876
- }
877
- default :
878
- AT_ERROR (" _th_eig not supported on CPUType for " , dispatch_scalar_type);
879
- }
880
- return std::tuple<Tensor, Tensor>(res1, res2);
881
- }
882
835
Tensor & _th_potri_out (Tensor & output, const Tensor & self, bool upper) {
883
836
// DeviceGuard omitted
884
837
auto dispatch_scalar_type = infer_scalar_type (self);
0 commit comments