Skip to content

Commit

Permalink
comment out lines
Browse files Browse the repository at this point in the history
  • Loading branch information
kojix2 committed Sep 27, 2023
1 parent c418f51 commit f328629
Showing 1 changed file with 24 additions and 24 deletions.
48 changes: 24 additions & 24 deletions ext/candle/src/lib.rs
Original file line number Diff line number Diff line change
Expand Up @@ -419,29 +419,29 @@ impl PyTensor {
Ok(PyTensor(self.0.min_keepdim(dim).map_err(wrap_err)?))
}

fn eq(&self, rhs: &Self) -> PyResult<Self> {
Ok(PyTensor(self.0.eq(rhs).map_err(wrap_err)?))
}
// fn eq(&self, rhs: &Self) -> PyResult<Self> {
// Ok(PyTensor(self.0.eq(rhs).map_err(wrap_err)?))
// }

fn ne(&self, rhs: &Self) -> PyResult<Self> {
Ok(PyTensor(self.0.ne(rhs).map_err(wrap_err)?))
}
// fn ne(&self, rhs: &Self) -> PyResult<Self> {
// Ok(PyTensor(self.0.ne(rhs).map_err(wrap_err)?))
// }

fn lt(&self, rhs: &Self) -> PyResult<Self> {
Ok(PyTensor(self.0.lt(rhs).map_err(wrap_err)?))
}
// fn lt(&self, rhs: &Self) -> PyResult<Self> {
// Ok(PyTensor(self.0.lt(rhs).map_err(wrap_err)?))
// }

fn gt(&self, rhs: &Self) -> PyResult<Self> {
Ok(PyTensor(self.0.gt(rhs).map_err(wrap_err)?))
}
// fn gt(&self, rhs: &Self) -> PyResult<Self> {
// Ok(PyTensor(self.0.gt(rhs).map_err(wrap_err)?))
// }

fn ge(&self, rhs: &Self) -> PyResult<Self> {
Ok(PyTensor(self.0.ge(rhs).map_err(wrap_err)?))
}
// fn ge(&self, rhs: &Self) -> PyResult<Self> {
// Ok(PyTensor(self.0.ge(rhs).map_err(wrap_err)?))
// }

fn le(&self, rhs: &Self) -> PyResult<Self> {
Ok(PyTensor(self.0.le(rhs).map_err(wrap_err)?))
}
// fn le(&self, rhs: &Self) -> PyResult<Self> {
// Ok(PyTensor(self.0.le(rhs).map_err(wrap_err)?))
// }

/// Returns the sum of the tensor.
/// &RETURNS&: Tensor
Expand Down Expand Up @@ -734,12 +734,12 @@ fn init(ruby: &Ruby) -> PyResult<()> {
rb_tensor.define_method("argmin_keepdim", method!(PyTensor::argmin_keepdim, 1))?;
rb_tensor.define_method("max_keepdim", method!(PyTensor::max_keepdim, 1))?;
rb_tensor.define_method("min_keepdim", method!(PyTensor::min_keepdim, 1))?;
rb_tensor.define_method("eq", method!(PyTensor::eq, 1))?;
rb_tensor.define_method("ne", method!(PyTensor::ne, 1))?;
rb_tensor.define_method("lt", method!(PyTensor::lt, 1))?;
rb_tensor.define_method("gt", method!(PyTensor::gt, 1))?;
rb_tensor.define_method("ge", method!(PyTensor::ge, 1))?;
rb_tensor.define_method("le", method!(PyTensor::le, 1))?;
// rb_tensor.define_method("eq", method!(PyTensor::eq, 1))?;
// rb_tensor.define_method("ne", method!(PyTensor::ne, 1))?;
// rb_tensor.define_method("lt", method!(PyTensor::lt, 1))?;
// rb_tensor.define_method("gt", method!(PyTensor::gt, 1))?;
// rb_tensor.define_method("ge", method!(PyTensor::ge, 1))?;
// rb_tensor.define_method("le", method!(PyTensor::le, 1))?;
rb_tensor.define_method("sum_all", method!(PyTensor::sum_all, 0))?;
rb_tensor.define_method("mean_all", method!(PyTensor::mean_all, 0))?;
rb_tensor.define_method("flatten_from", method!(PyTensor::flatten_from, 1))?;
Expand Down

0 comments on commit f328629

Please sign in to comment.