ttest_cost=read.csv("D:/R/cost_wh.csv") x<-ttest_cost$x t.test(x, alternative = c("two.sided"),mu = 2000,paired = FALSE,var.equal = FALSE,conf.level = 0.95)修改代码
时间: 2024-03-02 19:52:09 浏览: 22
根据您的需求,我修改了您的代码如下:
```
ttest_cost <- read.csv("D:/R/cost_wh.csv")
x <- ttest_cost$cost
t.test(x, alternative = "two.sided", mu = 2000, paired = FALSE, var.equal = FALSE, conf.level = 0.95)
```
我做出的修改包括:
1. 将 `ttest_cost$x` 改为 `ttest_cost$cost`,因为从 csv 文件中读取的数据应该是存储在 `cost` 列中的。
2. 把 `alternative = c("two.sided")` 改为 `alternative = "two.sided"`,因为 `alternative` 参数只需要一个字符串值,而不是一个字符向量。
3. 将 `paired = T` 改为 `paired = FALSE`,因为您的数据集中似乎没有配对的观测值。
希望这些修改能够帮助您解决问题。如果您还有其他问题,请随时问我。
相关问题
t_test_results = [] for i, column in enumerate(data1_norm.columns[:-1]): t, p = stats.ttest_ind(data1_norm[column][data1_norm.group == 1], data1_norm[column][data1_norm.group == 2]) t_test_results.append((column, t, p)) t_test_results = pd.DataFrame(t_test_results, columns=['variable', 't_value', 'p_value']) significant_results = t_test_results[t_test_results.p_value < 0.05]这个是代码,AttributeError: 'DataFrame' object has no attribute 'append'. Did you mean: '_append'?这个是错误显示怎么改正
这个错误提示说DataFrame对象没有append属性,可以改成使用concat方法将结果添加到新的DataFrame中,代码如下:
```
t_test_results = []
for i, column in enumerate(data1_norm.columns[:-1]):
t, p = stats.ttest_ind(data1_norm[column][data1_norm.group == 1], data1_norm[column][data1_norm.group == 2])
t_test_results.append(pd.DataFrame({'variable': column, 't_value': t, 'p_value': p}, index=[i]))
t_test_results = pd.concat(t_test_results, ignore_index=True)
significant_results = t_test_results[t_test_results.p_value < 0.05]
```
这样就可以将每次循环得到的结果按行添加到一个列表中,最后使用concat方法将列表中的DataFrame对象合并成一个DataFrame对象。
帮我把这段代码从tensorflow框架改成pytorch框架: import tensorflow as tf import os import numpy as np import matplotlib.pyplot as plt os.environ["CUDA_VISIBLE_DEVICES"] = "0" base_dir = 'E:/direction/datasetsall/' train_dir = os.path.join(base_dir, 'train_img/') validation_dir = os.path.join(base_dir, 'val_img/') train_cats_dir = os.path.join(train_dir, 'down') train_dogs_dir = os.path.join(train_dir, 'up') validation_cats_dir = os.path.join(validation_dir, 'down') validation_dogs_dir = os.path.join(validation_dir, 'up') batch_size = 64 epochs = 50 IMG_HEIGHT = 128 IMG_WIDTH = 128 num_cats_tr = len(os.listdir(train_cats_dir)) num_dogs_tr = len(os.listdir(train_dogs_dir)) num_cats_val = len(os.listdir(validation_cats_dir)) num_dogs_val = len(os.listdir(validation_dogs_dir)) total_train = num_cats_tr + num_dogs_tr total_val = num_cats_val + num_dogs_val train_image_generator = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1. / 255) validation_image_generator = tf.keras.preprocessing.image.ImageDataGenerator(rescale=1. / 255) train_data_gen = train_image_generator.flow_from_directory(batch_size=batch_size, directory=train_dir, shuffle=True, target_size=(IMG_HEIGHT, IMG_WIDTH), class_mode='categorical') val_data_gen = validation_image_generator.flow_from_directory(batch_size=batch_size, directory=validation_dir, target_size=(IMG_HEIGHT, IMG_WIDTH), class_mode='categorical') sample_training_images, _ = next(train_data_gen) model = tf.keras.models.Sequential([ tf.keras.layers.Conv2D(16, 3, padding='same', activation='relu', input_shape=(IMG_HEIGHT, IMG_WIDTH, 3)), tf.keras.layers.MaxPooling2D(), tf.keras.layers.Conv2D(32, 3, padding='same', activation='relu'), tf.keras.layers.MaxPooling2D(), tf.keras.layers.Conv2D(64, 3, padding='same', activation='relu'), tf.keras.layers.MaxPooling2D(), tf.keras.layers.Flatten(), tf.keras.layers.Dense(256, activation='relu'), tf.keras.layers.Dense(2, activation='softmax') ]) model.compile(optimizer='adam', loss=tf.keras.losses.BinaryCrossentropy(from_logits=True), metrics=['accuracy']) model.summary() history = model.fit_generator( train_data_gen, steps_per_epoch=total_train // batch_size, epochs=epochs, validation_data=val_data_gen, validation_steps=total_val // batch_size ) # 可视化训练结果 acc = history.history['accuracy'] val_acc = history.history['val_accuracy'] loss = history.history['loss'] val_loss = history.history['val_loss'] epochs_range = range(epochs) model.save("./model/timo_classification_128_maxPool2D_dense256.h5")
import torch import torch.nn as nn import torch.optim as optim from torch.utils.data import DataLoader from torchvision import datasets, transforms import os BATCH_SIZE = 64 EPOCHS = 50 IMG_HEIGHT = 128 IMG_WIDTH = 128 train_transforms = transforms.Compose([ transforms.Resize((IMG_HEIGHT,IMG_WIDTH)), transforms.RandomHorizontalFlip(), transforms.ToTensor(), transforms.Normalize([0.5,0.5,0.5], [0.5,0.5,0.5])]) test_transforms = transforms.Compose([ transforms.Resize((IMG_HEIGHT,IMG_WIDTH)), transforms.ToTensor(), transforms.Normalize([0.5,0.5,0.5], [0.5,0.5,0.5])]) base_dir = 'E:/direction/datasetsall/' train_dir = os.path.join(base_dir, 'train_img/') validation_dir = os.path.join(base_dir, 'val_img/') train_cats_dir = os.path.join(train_dir, 'down') train_dogs_dir = os.path.join(train_dir, 'up') validation_cats_dir = os.path.join(validation_dir, 'down') validation_dogs_dir = os.path.join(validation_dir, 'up') train_dataset = datasets.ImageFolder(train_dir, transform=train_transforms) train_loader = DataLoader(train_dataset, batch_size=BATCH_SIZE, shuffle=True) test_dataset = datasets.ImageFolder(validation_dir, transform=test_transforms) test_loader = DataLoader(test_dataset, batch_size=BATCH_SIZE, shuffle=False) device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu") model = nn.Sequential( nn.Conv2d(3, 16, 3, padding=1), nn.ReLU(), nn.MaxPool2d(2), nn.Conv2d(16, 32, 3, padding=1), nn.ReLU(), nn.MaxPool2d(2), nn.Conv2d(32, 64, 3, padding=1), nn.ReLU(), nn.MaxPool2d(2), nn.Flatten(), nn.Linear(64 * (IMG_HEIGHT // 8) * (IMG_WIDTH // 8), 256), nn.ReLU(), nn.Linear(256, 2), nn.Softmax(dim=1) ) model.to(device) criterion = nn.CrossEntropyLoss() optimizer = optim.Adam(model.parameters(), lr=0.001) for epoch in range(EPOCHS): train_loss = 0.0 train_acc = 0.0 model.train() for images, labels in train_loader: images = images.to(device) labels = labels.to(device) optimizer.zero_grad() outputs = model(images) loss = criterion(outputs, labels) loss.backward() optimizer.step() train_loss += loss.item() * images.size(0) _, preds = torch.max(outputs, 1) train_acc += torch.sum(preds == labels.data) train_loss = train_loss / len(train_loader.dataset) train_acc = train_acc / len(train_loader.dataset) print('Epoch: {} \tTraining Loss: {:.6f} \tTraining Accuracy: {:.6f}'.format(epoch+1, train_loss,train_acc)) with torch.no_grad(): test_loss = 0.0 test_acc = 0.0 model.eval() for images, labels in test_loader: images = images.to(device) labels = labels.to(device) outputs = model(images) loss = criterion(outputs, labels) test_loss += loss.item() * images.size(0) _, preds = torch.max(outputs, 1) test_acc += torch.sum(preds == labels.data) test_loss = test_loss / len(test_loader.dataset) test_acc = test_acc / len(test_loader.dataset) print('Test Loss: {:.6f} \tTest Accuracy: {:.6f}'.format(test_loss,test_acc))