[c-nsp] GEIP+ high CPU
Amol Sapkal
amolsapkal at gmail.com
Mon Dec 20 06:43:47 EST 2004
Refer this link:
http://www.cisco.com/en/US/products/sw/iosswrel/ps1828/products_tech_note09186a00800a65d0.shtml#showproccpu
On Mon, 20 Dec 2004 13:28:47 +0200, M.Palis <security at cytanet.com.cy> wrote:
> Here is the config. How did you understand that is interrupted switched?
>
> interface GigabitEthernet1/0/0
> bandwidth 10000000
> ip address x.x.x.x.x. no ip redirects
> no ip proxy-arp
> ip ospf message-digest-key 5 md5 7 xxxxxxxxxxxxxx
> no ip mroute-cache
> load-interval 30
> negotiation auto
> no cdp enable
> standby 40 ip x.x.x.x
> standby 40 priority 120
> standby 40 preempt
> !
> ----- Original Message -----
> From: "Amol Sapkal" <amolsapkal at gmail.com>
> To: "M.Palis" <security at cytanet.com.cy>
> Cc: <cisco-nsp at puck.nether.net>
> Sent: Monday, December 20, 2004 12:47 PM
> Subject: Re: [c-nsp] GEIP+ high CPU
>
> > Hi,
> >
> > Your process utilization output shows that all your traffic is
> > Interrupt switched (85% of 85%). I am not sure of this, but I think
> > lan interfaces (gig/fast/ethernet) should not be using interrupt
> > switching.
> >
> > Can you paste the relevant config of the gig interface?
> >
> >
> >
> > Regds,
> > Amol
> >
> >
> >
> > On Mon, 20 Dec 2004 11:06:30 +0200, M.Palis <security at cytanet.com.cy>
> > wrote:
> >> Hello all
> >> We are facing a high CPU utilization on a GEIP+ (avarage 80-90%). Below
> >> is
> >> the output of the show interface and sh contr vip 1 proc cpu which does
> >> not
> >> show which process causes the high CPU and why. I enable cache flow to
> >> see
> >> the type of traffic that passes through the GEIP+ but it seems that
> >> traffic
> >> is normal.
> >>
> >> Can you suggest something that will figure out what is the cause of high
> >> CPU
> >> utilization?
> >>
> >> GigabitEthernet1/0/0 is up, line protocol is up
> >> Hardware is cyBus GigabitEthernet Interface, address is 000b.60fb.6820
> >> (bia 000b.60fb.6820)
> >> Internet address is x.x.x.x.
> >> MTU 1500 bytes, BW 10000000 Kbit, DLY 10 usec,
> >> reliability 255/255, txload 5/255, rxload 2/255
> >> Encapsulation ARPA, loopback not set
> >> Keepalive set (10 sec)
> >> Full Duplex, 1000Mbps, Auto-negotiation,
> >> output flow-control is on, input flow-control is on
> >> Full-duplex, 1000Mb/s, link type is auto, media type is
> >> output flow-control is on, input flow-control is on
> >> ARP type: ARPA, ARP Timeout 04:00:00
> >> Last input 00:00:00, output 00:00:00, output hang never
> >> Last clearing of "show interface" counters never
> >> Input queue: 0/75/24425/167 (size/max/drops/flushes); Total output
> >> drops:
> >> 500
> >> Queueing strategy: fifo
> >> Output queue: 0/40 (size/max)
> >> 30 second input rate 99831000 bits/sec, 42356 packets/sec
> >> 30 second output rate 232347000 bits/sec, 44137 packets/sec
> >> 113608673126 packets input, 35803991154611 bytes, 0 no buffer
> >> Received 7049101 broadcasts (916211 IP multicast)
> >> 0 runts, 0 giants, 412 throttles
> >> 0 input errors, 0 CRC, 0 frame, 235891035 overrun, 179729695 ignored
> >> 0 watchdog, 0 multicast, 0 pause input
> >> 110887072498 packets output, 68984898771503 bytes, 0 underruns
> >> 0 output errors, 0 collisions, 2 interface resets
> >> 0 babbles, 0 late collision, 0 deferred
> >> 2 lost carrier, 0 no carrier, 0 PAUSE output
> >> 0 output buffer failures, 0 output buffers swapped out
> >>
> >> sh contr vip 1 proc cpu
> >> show proc cpu from Slot 1:
> >>
> >> CPU utilization for five seconds: 85%/85%; one minute: 86%; five minutes:
> >> 86%
> >> PID Runtime(ms) Invoked uSecs 5Sec 1Min 5Min TTY Process
> >> 1 0 1 0 0.00% 0.00% 0.00% 0 Chunk
> >> Manager
> >> 2 251048 537500 467 0.00% 0.00% 0.00% 0 Load Meter
> >> 3 7002796 4876298 1436 0.00% 0.00% 0.00% 0 CEF process
> >> 4 70565776 3054576 23101 0.00% 0.14% 0.14% 0 Check heaps
> >> 5 0 2 0 0.00% 0.00% 0.00% 0 Pool
> >> Manager
> >> 6 0 1 0 0.00% 0.00% 0.00% 0 Timers
> >> 7 0 1 0 0.00% 0.00% 0.00% 0 Serial
> >> Backgroun
> >> 8 10944 44781 244 0.00% 0.00% 0.00% 0 IPC Dynamic
> >> Cach
> >> 9 468876 190192 2465 0.00% 0.00% 0.00% 0 CEF Scanner
> >> 10 0 1 0 0.00% 0.00% 0.00% 0 IPC
> >> BackPressure
> >> 11 692964 2675813 258 0.00% 0.00% 0.00% 0 IPC
> >> Periodic
> >> Tim
> >> 12 540488 2679819 201 0.00% 0.00% 0.00% 0 IPC
> >> Deferred
> >> Por
> >> 13 60196 27093 2221 0.00% 0.00% 0.00% 0 IPC Seat
> >> Manager
> >> 14 0 1 0 0.00% 0.00% 0.00% 0 SERIAL
> >> A'detect
> >> 15 0 1 0 0.00% 0.00% 0.00% 0 Critical
> >> Bkgnd
> >> 16 1825468 350873 5202 0.00% 0.00% 0.00% 0 Net
> >> Background
> >> 17 0 6 0 0.00% 0.00% 0.00% 0 Logger
> >> 18 1065056 2675856 398 0.00% 0.00% 0.00% 0 TTY
> >> Background
> >> 19 6532620 2675467 2441 0.00% 0.00% 0.00% 0 Per-Second
> >> Jobs
> >> 20 6679672 44771 149199 0.00% 0.00% 0.00% 0 Per-minute
> >> Jobs
> >> 21 0 1 0 0.00% 0.00% 0.00% 0 CSP Timer
> >> 22 0 1 0 0.00% 0.00% 0.00% 0 SONET alarm
> >> time
> >> 23 0 1 0 0.00% 0.00% 0.00% 0 Hawkeye
> >> Backgrou
> >> 24 0 1 0 0.00% 0.00% 0.00% 0 VIP Encap
> >> IPC
> >> Ba
> >> 25 0 1 0 0.00% 0.00% 0.00% 0 MLP Input
> >> 26 12 1 12000 0.00% 0.00% 0.00% 0 IP Flow LC
> >> Backg
> >> 27 44964204 266488100 168 0.00% 0.00% 0.00% 0 VIP MEMD
> >> buffer
> >> 28 0 1 0 0.00% 0.00% 0.00% 0 AAA
> >> Dictionary R
> >> 29 0 2 0 0.00% 0.00% 0.00% 0 IP Hdr Comp
> >> Proc
> >> 30 9387952 26219499 358 0.00% 0.00% 0.00% 0 MDFS MFIB
> >> Proces
> >> 31 1018112 1677 607103 0.00% 0.00% 0.00% 0 TurboACL
> >> 32 47172612 26504344 1779 0.00% 0.01% 0.00% 0 CEF LC IPC
> >> Backg
> >> 33 10743144 3454406 3109 0.00% 0.00% 0.00% 0 CEF LC
> >> Stats
> >> 34 0 4 0 0.00% 0.00% 0.00% 0 CEF MQC IPC
> >> Back
> >> 35 0 1 0 0.00% 0.00% 0.00% 0 TFIB LC
> >> cleanup
> >> 36 0 1 0 0.00% 0.00% 0.00% 0 Any
> >> Transport
> >> ov
> >> 37 0 1 0 0.00% 0.00% 0.00% 0 MDFS LC
> >> Process
> >> 38 0 1 0 0.00% 0.00% 0.00% 0 LI LC
> >> Messaging
> >> 39 143852 24419 5890 0.00% 0.00% 0.00% 0 Clock
> >> Client
> >> 40 84956 537101 158 0.00% 0.00% 0.00% 0 DBUS
> >> Console
> >> 41 0 1 0 0.00% 0.00% 0.00% 0 Net Input
> >> 42 249052 537499 463 0.00% 0.00% 0.00% 0 Compute
> >> load
> >> avg
> >> 43 0 1 0 0.00% 0.00% 0.00% 0 IP Flow
> >> Backgrou
> >> 44 120 27 4444 0.00% 0.00% 0.00% 1
> >> console_rpc_serv
> >>
> >> _______________________________________________
> >> cisco-nsp mailing list cisco-nsp at puck.nether.net
> >> https://puck.nether.net/mailman/listinfo/cisco-nsp
> >> archive at http://puck.nether.net/pipermail/cisco-nsp/
> >>
> >
> >
> > --
> > Warm Regds,
> >
> > Amol Sapkal
> >
> > --------------------------------------------------------------------
> > An eye for an eye makes the whole world blind
> > - Mahatma Gandhi
> > --------------------------------------------------------------------
>
>
--
Warm Regds,
Amol Sapkal
--------------------------------------------------------------------
An eye for an eye makes the whole world blind
- Mahatma Gandhi
--------------------------------------------------------------------
More information about the cisco-nsp
mailing list