util-linux/sys-utils/lscpu.c

882 lines
20 KiB
C
Raw Normal View History

/*
* lscpu - CPU architecture information helper
*
* Copyright (C) 2008 Cai Qian <qcai@redhat.com>
* Copyright (C) 2008 Karel Zak <kzak@redhat.com>
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation, either version 3 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program. If not, see <http://www.gnu.org/licenses/>.
*/
#include <ctype.h>
#include <dirent.h>
#include <err.h>
#include <errno.h>
#include <fcntl.h>
#include <getopt.h>
#include <stdio.h>
#include <stdlib.h>
#include <string.h>
#include <sys/utsname.h>
#include <unistd.h>
#include <stdarg.h>
#include "cpuset.h"
#include "nls.h"
#define CACHE_MAX 100
/* /sys paths */
#define _PATH_SYS_SYSTEM "/sys/devices/system"
#define _PATH_SYS_CPU _PATH_SYS_SYSTEM "/cpu"
#define _PATH_PROC_XEN "/proc/xen"
#define _PATH_PROC_XENCAP _PATH_PROC_XEN "/capabilities"
#define _PATH_PROC_CPUINFO "/proc/cpuinfo"
#define _PATH_PROC_PCIDEVS "/proc/bus/pci/devices"
/* virtualization types */
enum {
VIRT_NONE = 0,
VIRT_PARA,
VIRT_FULL
};
const char *virt_types[] = {
[VIRT_NONE] = N_("none"),
[VIRT_PARA] = N_("para"),
[VIRT_FULL] = N_("full")
};
/* hypervisor vendors */
enum {
HYPER_NONE = 0,
HYPER_XEN,
HYPER_KVM,
HYPER_MSHV
};
const char *hv_vendors[] = {
[HYPER_NONE] = NULL,
[HYPER_XEN] = "Xen",
[HYPER_KVM] = "KVM",
[HYPER_MSHV] = "Microsoft"
};
/* CPU modes (bits) */
enum {
MODE_REAL = (1 << 1),
MODE_TRANSPARENT = (1 << 2),
MODE_LONG = (1 << 3)
};
/* cache(s) description */
struct cpu_cache {
char *name;
char *size;
int nsharedmaps;
cpu_set_t **sharedmaps;
};
/* global description */
struct lscpu_desc {
char *arch;
char *vendor;
char *family;
char *model;
char *virtflag; /* virtualization flag (vmx, svm) */
int hyper; /* hypervisor vendor ID */
int virtype; /* VIRT_PARA|FULL|NONE ? */
char *mhz;
char *stepping;
char *flags;
int mode; /* rm, lm or/and tm */
int ncpus; /* number of CPUs */
int nnodes; /* number of NUMA modes */
cpu_set_t **nodemaps; /* array with NUMA nodes */
/* sockets -- based on core_siblings (internal kernel map of cpuX's
* hardware threads within the same physical_package_id (socket)) */
int nsockets; /* number of all sockets */
cpu_set_t **socketmaps; /* unique core_siblings */
/* cores -- based on thread_siblings (internel kernel map of cpuX's
* hardware threads within the same core as cpuX) */
int ncores; /* number of all cores */
cpu_set_t **coremaps; /* unique thread_siblings */
int nthreads; /* number of threads */
int ncaches;
struct cpu_cache *caches;
};
static size_t sysrootlen;
static char pathbuf[PATH_MAX];
static int maxcpus; /* size in bits of kernel cpu mask */
static FILE *path_fopen(const char *mode, int exit_on_err, const char *path, ...)
__attribute__ ((__format__ (__printf__, 3, 4)));
static void path_getstr(char *result, size_t len, const char *path, ...)
__attribute__ ((__format__ (__printf__, 3, 4)));
static int path_getnum(const char *path, ...)
__attribute__ ((__format__ (__printf__, 1, 2)));
static int path_exist(const char *path, ...)
__attribute__ ((__format__ (__printf__, 1, 2)));
static cpu_set_t *path_cpuset(const char *path, ...)
__attribute__ ((__format__ (__printf__, 1, 2)));
static const char *
path_vcreate(const char *path, va_list ap)
{
if (sysrootlen)
vsnprintf(pathbuf + sysrootlen,
sizeof(pathbuf) - sysrootlen, path, ap);
else
vsnprintf(pathbuf, sizeof(pathbuf), path, ap);
return pathbuf;
}
static FILE *
path_vfopen(const char *mode, int exit_on_error, const char *path, va_list ap)
{
FILE *f;
const char *p = path_vcreate(path, ap);
f = fopen(p, mode);
if (!f && exit_on_error)
err(EXIT_FAILURE, _("error: cannot open %s"), p);
return f;
}
static FILE *
path_fopen(const char *mode, int exit_on_error, const char *path, ...)
{
FILE *fd;
va_list ap;
va_start(ap, path);
fd = path_vfopen("r", exit_on_error, path, ap);
va_end(ap);
return fd;
}
static void
path_getstr(char *result, size_t len, const char *path, ...)
{
FILE *fd;
va_list ap;
va_start(ap, path);
fd = path_vfopen("r", 1, path, ap);
va_end(ap);
if (!fgets(result, len, fd))
err(EXIT_FAILURE, _("failed to read: %s"), pathbuf);
fclose(fd);
len = strlen(result);
if (result[len - 1] == '\n')
result[len - 1] = '\0';
}
static int
path_getnum(const char *path, ...)
{
FILE *fd;
va_list ap;
int result;
va_start(ap, path);
fd = path_vfopen("r", 1, path, ap);
va_end(ap);
if (fscanf(fd, "%d", &result) != 1) {
if (ferror(fd))
err(EXIT_FAILURE, _("failed to read: %s"), pathbuf);
else
errx(EXIT_FAILURE, _("parse error: %s"), pathbuf);
}
fclose(fd);
return result;
}
static int
path_exist(const char *path, ...)
{
va_list ap;
const char *p;
va_start(ap, path);
p = path_vcreate(path, ap);
va_end(ap);
return access(p, F_OK) == 0;
}
static char *
xstrdup(const char *str)
{
char *s = strdup(str);
if (!s)
err(EXIT_FAILURE, _("error: strdup failed"));
return s;
}
static cpu_set_t *
path_cpuset(const char *path, ...)
{
FILE *fd;
va_list ap;
cpu_set_t *set;
size_t setsize, len = maxcpus * 7;
char buf[len];
va_start(ap, path);
fd = path_vfopen("r", 1, path, ap);
va_end(ap);
if (!fgets(buf, len, fd))
err(EXIT_FAILURE, _("failed to read: %s"), pathbuf);
fclose(fd);
len = strlen(buf);
if (buf[len - 1] == '\n')
buf[len - 1] = '\0';
set = cpuset_alloc(maxcpus, &setsize, NULL);
if (!set)
err(EXIT_FAILURE, _("failed to callocate cpu set"));
if (cpumask_parse(buf, set, setsize))
errx(EXIT_FAILURE, _("failed to parse CPU mask %s"), buf);
return set;
}
/* Lookup a pattern and get the value from cpuinfo.
* Format is:
*
* "<pattern> : <key>"
*/
int lookup(char *line, char *pattern, char **value)
{
char *p, *v;
int len = strlen(pattern);
if (!*line)
return 0;
/* pattern */
if (strncmp(line, pattern, len))
return 0;
/* white spaces */
for (p = line + len; isspace(*p); p++);
/* separator */
if (*p != ':')
return 0;
/* white spaces */
for (++p; isspace(*p); p++);
/* value */
if (!*p)
return 0;
v = p;
/* end of value */
len = strlen(line) - 1;
for (p = line + len; isspace(*(p-1)); p--);
*p = '\0';
*value = xstrdup(v);
return 1;
}
static void
read_basicinfo(struct lscpu_desc *desc)
{
FILE *fp = path_fopen("r", 1, _PATH_PROC_CPUINFO);
char buf[BUFSIZ];
struct utsname utsbuf;
/* architecture */
if (uname(&utsbuf) == -1)
err(EXIT_FAILURE, _("error: uname failed"));
desc->arch = xstrdup(utsbuf.machine);
/* count CPU(s) */
while(path_exist(_PATH_SYS_SYSTEM "/cpu/cpu%d", desc->ncpus))
desc->ncpus++;
/* details */
while (fgets(buf, sizeof(buf), fp) != NULL) {
/* IA64 */
if (lookup(buf, "vendor", &desc->vendor)) ;
else if (lookup(buf, "vendor_id", &desc->vendor)) ;
/* IA64 */
else if (lookup(buf, "family", &desc->family)) ;
else if (lookup(buf, "cpu family", &desc->family)) ;
else if (lookup(buf, "model", &desc->model)) ;
else if (lookup(buf, "stepping", &desc->stepping)) ;
else if (lookup(buf, "cpu MHz", &desc->mhz)) ;
else if (lookup(buf, "flags", &desc->flags)) ;
else
continue;
}
if (desc->flags) {
snprintf(buf, sizeof(buf), " %s ", desc->flags);
if (strstr(buf, " svm "))
desc->virtflag = strdup("svm");
else if (strstr(buf, " vmx "))
desc->virtflag = strdup("vmx");
if (strstr(buf, " rm "))
desc->mode |= MODE_REAL;
if (strstr(buf, " tm "))
desc->mode |= MODE_TRANSPARENT;
if (strstr(buf, " lm "))
desc->mode |= MODE_LONG;
}
fclose(fp);
if (path_exist(_PATH_SYS_SYSTEM "/cpu/kernel_max"))
maxcpus = path_getnum(_PATH_SYS_SYSTEM "/cpu/kernel_max");
else if (!sysrootlen)
/* the root is '/' so we are working with data from the current kernel */
maxcpus = get_max_number_of_cpus();
else
/* we are reading some /sys snapshot instead of the real /sys,
* let's use any crazy number... */
maxcpus = desc->ncpus > 2048 ? desc->ncpus : 2048;
}
static int
has_pci_device(int vendor, int device)
{
FILE *f;
int num, fn, ven, dev;
int res = 1;
f = path_fopen("r", 0, _PATH_PROC_PCIDEVS);
if (!f)
return 0;
/* for more details about bus/pci/devices format see
* drivers/pci/proc.c in linux kernel
*/
while(fscanf(f, "%02x%02x\t%04x%04x\t%*[^\n]",
&num, &fn, &ven, &dev) == 4) {
if (ven == vendor && dev == device)
goto found;
}
res = 0;
found:
fclose(f);
return res;
}
#if defined(__x86_64__) || defined(__i386__)
/*
* This CPUID leaf returns the information about the hypervisor.
* EAX : maximum input value for CPUID supported by the hypervisor.
* EBX, ECX, EDX : Hypervisor vendor ID signature. E.g. VMwareVMware.
*/
#define HYPERVISOR_INFO_LEAF 0x40000000
static inline void
cpuid(unsigned int op, unsigned int *eax, unsigned int *ebx,
unsigned int *ecx, unsigned int *edx)
{
__asm__(
#if defined(__PIC__) && defined(__i386__)
/* x86 PIC cannot clobber ebx -- gcc bitches */
"pushl %%ebx;"
"cpuid;"
"movl %%ebx, %%esi;"
"popl %%ebx;"
: "=S" (*ebx),
#else
"cpuid;"
: "=b" (*ebx),
#endif
"=a" (*eax),
"=c" (*ecx),
"=d" (*edx)
: "1" (op), "c"(0));
}
static void
read_hypervisor_cpuid(struct lscpu_desc *desc)
{
unsigned int eax = 0, ebx = 0, ecx = 0, edx = 0;
char hyper_vendor_id[13];
memset(hyper_vendor_id, 0, sizeof(hyper_vendor_id));
cpuid(HYPERVISOR_INFO_LEAF, &eax, &ebx, &ecx, &edx);
memcpy(hyper_vendor_id + 0, &ebx, 4);
memcpy(hyper_vendor_id + 4, &ecx, 4);
memcpy(hyper_vendor_id + 8, &edx, 4);
hyper_vendor_id[12] = '\0';
if (!hyper_vendor_id[0])
return;
if (!strncmp("XenVMMXenVMM", hyper_vendor_id, 12))
desc->hyper = HYPER_XEN;
else if (!strncmp("KVMKVMKVM", hyper_vendor_id, 9))
desc->hyper = HYPER_KVM;
else if (!strncmp("Microsoft Hv", hyper_vendor_id, 12))
desc->hyper = HYPER_MSHV;
}
#else /* ! __x86_64__ */
static void
read_hypervisor_cpuid(struct lscpu_desc *desc)
{
}
#endif
static void
read_hypervisor(struct lscpu_desc *desc)
{
read_hypervisor_cpuid(desc);
if (desc->hyper)
/* hvm */
desc->virtype = VIRT_FULL;
else if (path_exist(_PATH_PROC_XEN)) {
/* Xen para-virt or dom0 */
FILE *fd = path_fopen("r", 0, _PATH_PROC_XENCAP);
int dom0 = 0;
if (fd) {
char buf[256];
if (fscanf(fd, "%s", buf) == 1 &&
!strcmp(buf, "control_d"))
dom0 = 1;
fclose(fd);
}
desc->virtype = dom0 ? VIRT_NONE : VIRT_PARA;
desc->hyper = HYPER_XEN;
} else if (has_pci_device(0x5853, 0x0001)) {
/* Xen full-virt on non-x86_64 */
desc->hyper = HYPER_XEN;
desc->virtype = VIRT_FULL;
}
}
/* add @set to the @ary, unnecesary set is deallocated. */
static int add_cpuset_to_array(cpu_set_t **ary, int *items, cpu_set_t *set)
{
int i;
size_t setsize = CPU_ALLOC_SIZE(maxcpus);
if (!ary)
return -1;
for (i = 0; i < *items; i++) {
if (CPU_EQUAL_S(setsize, set, ary[i]))
break;
}
if (i == *items) {
ary[*items] = set;
++*items;
return 0;
}
CPU_FREE(set);
return 1;
}
static void
read_topology(struct lscpu_desc *desc, int num)
{
cpu_set_t *thread_siblings, *core_siblings;
if (!path_exist(_PATH_SYS_CPU "/cpu%d/topology/thread_siblings", num))
return;
thread_siblings = path_cpuset(_PATH_SYS_CPU
"/cpu%d/topology/thread_siblings", num);
core_siblings = path_cpuset(_PATH_SYS_CPU
"/cpu%d/topology/core_siblings", num);
if (num == 0) {
int ncores, nsockets, nthreads;
size_t setsize = CPU_ALLOC_SIZE(maxcpus);
/* threads within one core */
nthreads = CPU_COUNT_S(setsize, thread_siblings);
/* cores within one socket */
ncores = CPU_COUNT_S(setsize, core_siblings) / nthreads;
/* number of sockets */
nsockets = desc->ncpus / nthreads / ncores;
/* all threads */
desc->nthreads = nsockets * ncores * nthreads;
desc->socketmaps = calloc(nsockets, sizeof(cpu_set_t *));
if (!desc->socketmaps)
err(EXIT_FAILURE, _("error: calloc failed"));
desc->coremaps = calloc(ncores * nsockets, sizeof(cpu_set_t *));
if (!desc->coremaps)
err(EXIT_FAILURE, _("error: calloc failed"));
}
add_cpuset_to_array(desc->socketmaps, &desc->nsockets, core_siblings);
add_cpuset_to_array(desc->coremaps, &desc->ncores, thread_siblings);
}
static int
cachecmp(const void *a, const void *b)
{
struct cpu_cache *c1 = (struct cpu_cache *) a;
struct cpu_cache *c2 = (struct cpu_cache *) b;
return strcmp(c2->name, c1->name);
}
static void
read_cache(struct lscpu_desc *desc, int num)
{
char buf[256];
int i;
if (num == 0) {
while(path_exist(_PATH_SYS_SYSTEM "/cpu/cpu%d/cache/index%d",
num, desc->ncaches))
desc->ncaches++;
if (!desc->ncaches)
return;
desc->caches = calloc(desc->ncaches, sizeof(*desc->caches));
if (!desc->caches)
err(EXIT_FAILURE, _("calloc failed"));
}
for (i = 0; i < desc->ncaches; i++) {
struct cpu_cache *ca = &desc->caches[i];
cpu_set_t *map;
if (!ca->name) {
int type, level;
/* cache type */
path_getstr(buf, sizeof(buf),
_PATH_SYS_CPU "/cpu%d/cache/index%d/type",
num, i);
if (!strcmp(buf, "Data"))
type = 'd';
else if (!strcmp(buf, "Instruction"))
type = 'i';
else
type = 0;
/* cache level */
level = path_getnum(_PATH_SYS_CPU "/cpu%d/cache/index%d/level",
num, i);
if (type)
snprintf(buf, sizeof(buf), "L%d%c", level, type);
else
snprintf(buf, sizeof(buf), "L%d", level);
ca->name = xstrdup(buf);
/* cache size */
path_getstr(buf, sizeof(buf),
_PATH_SYS_CPU "/cpu%d/cache/index%d/size",
num, i);
ca->size = xstrdup(buf);
}
/* information about how CPUs share different caches */
map = path_cpuset(_PATH_SYS_CPU "/cpu%d/cache/index%d/shared_cpu_map",
num, i);
if (!ca->sharedmaps) {
ca->sharedmaps = calloc(desc->ncpus, sizeof(cpu_set_t *));
if (!ca->sharedmaps)
err(EXIT_FAILURE, _("error: calloc failed"));
}
add_cpuset_to_array(ca->sharedmaps, &ca->nsharedmaps, map);
}
}
static void
read_nodes(struct lscpu_desc *desc)
{
int i;
/* number of NUMA node */
while (path_exist(_PATH_SYS_SYSTEM "/node/node%d", desc->nnodes))
desc->nnodes++;
if (!desc->nnodes)
return;
desc->nodemaps = calloc(desc->nnodes, sizeof(cpu_set_t *));
if (!desc->nodemaps)
err(EXIT_FAILURE, _("error: calloc failed"));
/* information about how nodes share different CPUs */
for (i = 0; i < desc->nnodes; i++)
desc->nodemaps[i] = path_cpuset(
_PATH_SYS_SYSTEM "/node/node%d/cpumap",
i);
}
static void
print_parsable(struct lscpu_desc *desc)
{
int i, j;
size_t setsize = CPU_ALLOC_SIZE(maxcpus);
printf(_(
"# The following is the parsable format, which can be fed to other\n"
"# programs. Each different item in every column has an unique ID\n"
"# starting from zero.\n"
"# CPU,Core,Socket,Node"));
if (desc->ncaches) {
/* separator between CPU topology and cache information */
putchar(',');
for (i = desc->ncaches - 1; i >= 0; i--)
printf(",%s", desc->caches[i].name);
}
putchar('\n');
for (i = 0; i < desc->ncpus; i++) {
/* #CPU */
printf("%d", i);
/* Core */
for (j = 0; j < desc->ncores; j++) {
if (CPU_ISSET_S(i, setsize, desc->coremaps[j])) {
printf(",%d", j);
break;
}
}
if (j == desc->ncores)
putchar(',');
/* Socket */
for (j = 0; j < desc->nsockets; j++) {
if (CPU_ISSET_S(i, setsize, desc->socketmaps[j])) {
printf(",%d", j);
break;
}
}
if (j == desc->nsockets)
putchar(',');
/* Nodes */
for (j = 0; j < desc->nnodes; j++) {
if (CPU_ISSET_S(i, setsize, desc->nodemaps[j])) {
printf(",%d", j);
break;
}
}
if (j == desc->nnodes)
putchar(',');
if (desc->ncaches)
putchar(',');
/* Caches */
for (j = desc->ncaches - 1; j >= 0; j--) {
struct cpu_cache *ca = &desc->caches[j];
int x;
for (x = 0; x < ca->nsharedmaps; x++) {
if (CPU_ISSET_S(i, setsize, ca->sharedmaps[x])) {
printf(",%d", x);
break;
}
}
if (x == ca->nsharedmaps)
putchar(',');
}
putchar('\n');
}
}
/* output formats "<key> <value>"*/
#define print_s(_key, _val) printf("%-23s%s\n", _key, _val)
#define print_n(_key, _val) printf("%-23s%d\n", _key, _val)
static void
print_readable(struct lscpu_desc *desc)
{
char buf[512];
int i;
print_s(_("Architecture:"), desc->arch);
if (desc->mode & (MODE_REAL | MODE_TRANSPARENT | MODE_LONG)) {
char buf[64], *p = buf;
if (desc->mode & MODE_REAL) {
strcpy(p, "16-bit, ");
p += 8;
}
if (desc->mode & MODE_TRANSPARENT) {
strcpy(p, "32-bit, ");
p += 8;
}
if (desc->mode & MODE_LONG) {
strcpy(p, "64-bit, ");
p += 8;
}
*(p - 2) = '\0';
print_s(_("CPU op-mode(s):"), buf);
}
print_n(_("CPU(s):"), desc->ncpus);
if (desc->nsockets) {
print_n(_("Thread(s) per core:"), desc->nthreads / desc->ncores);
print_n(_("Core(s) per socket:"), desc->ncores / desc->nsockets);
print_n(_("CPU socket(s):"), desc->nsockets);
}
if (desc->nnodes)
print_n(_("NUMA node(s):"), desc->nnodes);
if (desc->vendor)
print_s(_("Vendor ID:"), desc->vendor);
if (desc->family)
print_s(_("CPU family:"), desc->family);
if (desc->model)
print_s(_("Model:"), desc->model);
if (desc->stepping)
print_s(_("Stepping:"), desc->stepping);
if (desc->mhz)
print_s(_("CPU MHz:"), desc->mhz);
if (desc->virtflag) {
if (!strcmp(desc->virtflag, "svm"))
print_s(_("Virtualization:"), "AMD-V");
else if (!strcmp(desc->virtflag, "vmx"))
print_s(_("Virtualization:"), "VT-x");
}
if (desc->hyper) {
print_s(_("Hypervisor vendor:"), hv_vendors[desc->hyper]);
print_s(_("Virtualization type:"), virt_types[desc->virtype]);
}
if (desc->ncaches) {
char buf[512];
int i;
for (i = desc->ncaches - 1; i >= 0; i--) {
snprintf(buf, sizeof(buf),
_("%s cache:"), desc->caches[i].name);
print_s(buf, desc->caches[i].size);
}
}
if (desc->nnodes) {
size_t setbuflen = 7 * maxcpus;
char setbuf[setbuflen];
for (i = 0; i < desc->nnodes; i++) {
snprintf(buf, sizeof(buf), _("NUMA node%d CPU(s):"), i);
print_s(buf, cpulist_create(
setbuf, setbuflen,
desc->nodemaps[i],
CPU_ALLOC_SIZE(maxcpus)));
}
}
}
void usage(int rc)
{
printf(_("Usage: %s [option]\n"),
program_invocation_short_name);
puts(_( "CPU architecture information helper\n\n"
" -h, --help usage information\n"
" -p, --parse print out in parsable instead of printable format.\n"
" -s, --sysroot use the directory as a new system root.\n"));
exit(rc);
}
int main(int argc, char *argv[])
{
struct lscpu_desc _desc, *desc = &_desc;
int parsable = 0, c, i;
struct option longopts[] = {
{ "help", no_argument, 0, 'h' },
{ "parse", no_argument, 0, 'p' },
{ "sysroot", required_argument, 0, 's' },
{ NULL, 0, 0, 0 }
};
setlocale(LC_ALL, "");
bindtextdomain(PACKAGE, LOCALEDIR);
textdomain(PACKAGE);
while((c = getopt_long(argc, argv, "hps:", longopts, NULL)) != -1) {
switch (c) {
case 'h':
usage(EXIT_SUCCESS);
case 'p':
parsable = 1;
break;
case 's':
sysrootlen = strlen(optarg);
strncpy(pathbuf, optarg, sizeof(pathbuf));
pathbuf[sizeof(pathbuf) - 1] = '\0';
break;
default:
usage(EXIT_FAILURE);
}
}
memset(desc, 0, sizeof(*desc));
read_basicinfo(desc);
for (i = 0; i < desc->ncpus; i++) {
read_topology(desc, i);
read_cache(desc, i);
}
qsort(desc->caches, desc->ncaches, sizeof(struct cpu_cache), cachecmp);
read_nodes(desc);
read_hypervisor(desc);
/* Show time! */
if (parsable)
print_parsable(desc);
else
print_readable(desc);
return EXIT_SUCCESS;
}